Generalization Bounds of Deep Neural Networks With τ-Mixing Samples

IF 8.9 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE IEEE transactions on neural networks and learning systems Pub Date : 2025-01-22 DOI:10.1109/TNNLS.2025.3526235
Liyuan Liu;Yaohui Chen;Weifu Li;Yingjie Wang;Bin Gu;Feng Zheng;Hong Chen
{"title":"Generalization Bounds of Deep Neural Networks With τ-Mixing Samples","authors":"Liyuan Liu;Yaohui Chen;Weifu Li;Yingjie Wang;Bin Gu;Feng Zheng;Hong Chen","doi":"10.1109/TNNLS.2025.3526235","DOIUrl":null,"url":null,"abstract":"Deep neural networks (DNNs) have shown an astonishing ability to unlock the complicated relationships among the inputs and their responses. Along with empirical successes, some approximation analysis of DNNs has also been provided to understand their generalization performance. However, the existing analysis depends heavily on the independently identically distribution (i.i.d.) assumption of observations, which may be too ideal and often violated in real-world applications. To relax the i.i.d. assumption, this article develops the covering number-based concentration estimation to establish generalization bounds of DNNs with <inline-formula> <tex-math>$\\tau $ </tex-math></inline-formula>-mixing samples, where the dependency between samples is much general including <inline-formula> <tex-math>$\\alpha $ </tex-math></inline-formula>-mixing process as a special case. By assigning a specific parameter value to the <inline-formula> <tex-math>$\\tau $ </tex-math></inline-formula>-mixing process, our results are consistent with the existing convergence analysis under the i.i.d. case. Experiments on simulated data validate the theoretical findings.","PeriodicalId":13303,"journal":{"name":"IEEE transactions on neural networks and learning systems","volume":"36 8","pages":"14596-14610"},"PeriodicalIF":8.9000,"publicationDate":"2025-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on neural networks and learning systems","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10848487/","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Deep neural networks (DNNs) have shown an astonishing ability to unlock the complicated relationships among the inputs and their responses. Along with empirical successes, some approximation analysis of DNNs has also been provided to understand their generalization performance. However, the existing analysis depends heavily on the independently identically distribution (i.i.d.) assumption of observations, which may be too ideal and often violated in real-world applications. To relax the i.i.d. assumption, this article develops the covering number-based concentration estimation to establish generalization bounds of DNNs with $\tau $ -mixing samples, where the dependency between samples is much general including $\alpha $ -mixing process as a special case. By assigning a specific parameter value to the $\tau $ -mixing process, our results are consistent with the existing convergence analysis under the i.i.d. case. Experiments on simulated data validate the theoretical findings.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
$\tau$混合样本下深度神经网络的泛化界
深度神经网络(dnn)已经显示出惊人的能力,可以解开输入和响应之间的复杂关系。随着经验上的成功,一些dnn的近似分析也被提供来理解它们的泛化性能。然而,现有的分析在很大程度上依赖于观测值的独立同分布假设,这种假设可能过于理想,并且在实际应用中经常被违反。为了放宽i.i.d假设,本文发展了基于覆盖数的浓度估计来建立含有$\tau $ -混合样本的dnn的泛化边界,其中样本之间的依赖关系非常普遍,其中$\alpha $ -混合过程是一个特例。通过对$\tau $ -混合过程赋予特定的参数值,我们的结果与已有的i.i.d.情况下的收敛性分析一致。模拟数据实验验证了理论结果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
IEEE transactions on neural networks and learning systems
IEEE transactions on neural networks and learning systems COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE-COMPUTER SCIENCE, HARDWARE & ARCHITECTURE
CiteScore
23.80
自引率
9.60%
发文量
2102
审稿时长
3-8 weeks
期刊介绍: The focus of IEEE Transactions on Neural Networks and Learning Systems is to present scholarly articles discussing the theory, design, and applications of neural networks as well as other learning systems. The journal primarily highlights technical and scientific research in this domain.
期刊最新文献
NPSVC++: A Representation Learning Framework for Nonparallel Classifiers. Heuristic Knowledge-Driven Spatio-Temporal Forecasting via Multigraph. Robust Image-Based Visual Servoing Formation Control for Quadrotors Without Communication via Reinforcement Learning. Virtual Domain-Guided Cross-Modal Distillation With Multiview Correlation Awareness for Domain-Specific Multimodal Neural Machine Translation. Redundancy Removal and Knowledge Alignment-Based Personalized Federated Learning for Online Condition Monitoring
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1