首页 > 最新文献

Neurocomputing最新文献

英文 中文
Review of neural network model acceleration techniques based on FPGA platforms 基于 FPGA 平台的神经网络模型加速技术综述
IF 5.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-31 DOI: 10.1016/j.neucom.2024.128511

Neural network models, celebrated for their outstanding scalability and computational capabilities, have demonstrated remarkable performance across various fields such as vision, language, and multimodality. The rapid advancements in neural networks, fueled by the deep development of Internet technology and the increasing demand for intelligent edge devices, introduce new challenges, including significant model parameter sizes and increased storage pressures. In this context, Field-Programmable Gate Arrays (FPGA) emerge as a preferred platform for accelerating neural network models, thanks to their exceptional performance, energy efficiency, and the flexibility and scalability of the system. Building FPGA-based neural network systems necessitates bridging significant differences in objectives, methods, and design spaces between model design and hardware design. This review article adopts a comprehensive analytical framework to thoroughly explore multidimensional technological implementation strategies, encompassing optimizations at the algorithmic and hardware levels, as well as compiler optimization techniques. It focuses on methods for collaborative optimization between algorithms and hardware, identifies challenges in the collaborative design process, and proposes corresponding implementation strategies and key steps. Addressing various technological dimensions, the article provides in-depth technical analysis and discussion, aiming to offer valuable insights for research on optimizing and accelerating neural network models in edge computing environments.

神经网络模型以其出色的可扩展性和计算能力而著称,在视觉、语言和多模态等各个领域都表现出了卓越的性能。互联网技术的深入发展和对智能边缘设备日益增长的需求推动了神经网络的快速发展,同时也带来了新的挑战,包括巨大的模型参数规模和更大的存储压力。在这种情况下,现场可编程门阵列(FPGA)凭借其卓越的性能、能效以及系统的灵活性和可扩展性,成为加速神经网络模型的首选平台。构建基于 FPGA 的神经网络系统需要弥合模型设计和硬件设计在目标、方法和设计空间上的显著差异。本综述文章采用综合分析框架,深入探讨多维技术实现策略,包括算法和硬件层面的优化以及编译器优化技术。文章重点探讨了算法与硬件协同优化的方法,明确了协同设计过程中的挑战,并提出了相应的实施策略和关键步骤。文章针对不同的技术维度,进行了深入的技术分析和讨论,旨在为边缘计算环境下优化和加速神经网络模型的研究提供有价值的见解。
{"title":"Review of neural network model acceleration techniques based on FPGA platforms","authors":"","doi":"10.1016/j.neucom.2024.128511","DOIUrl":"10.1016/j.neucom.2024.128511","url":null,"abstract":"<div><p>Neural network models, celebrated for their outstanding scalability and computational capabilities, have demonstrated remarkable performance across various fields such as vision, language, and multimodality. The rapid advancements in neural networks, fueled by the deep development of Internet technology and the increasing demand for intelligent edge devices, introduce new challenges, including significant model parameter sizes and increased storage pressures. In this context, Field-Programmable Gate Arrays (FPGA) emerge as a preferred platform for accelerating neural network models, thanks to their exceptional performance, energy efficiency, and the flexibility and scalability of the system. Building FPGA-based neural network systems necessitates bridging significant differences in objectives, methods, and design spaces between model design and hardware design. This review article adopts a comprehensive analytical framework to thoroughly explore multidimensional technological implementation strategies, encompassing optimizations at the algorithmic and hardware levels, as well as compiler optimization techniques. It focuses on methods for collaborative optimization between algorithms and hardware, identifies challenges in the collaborative design process, and proposes corresponding implementation strategies and key steps. Addressing various technological dimensions, the article provides in-depth technical analysis and discussion, aiming to offer valuable insights for research on optimizing and accelerating neural network models in edge computing environments.</p></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":null,"pages":null},"PeriodicalIF":5.5,"publicationDate":"2024-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142168372","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Neural architecture search for image super-resolution: A review on the emerging state-of-the-art 图像超分辨率的神经架构搜索:最新技术综述
IF 5.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-31 DOI: 10.1016/j.neucom.2024.128481

Nowadays, complex and expensive neural architectures are seen by many as a way to improve the performance of existing models in image recognition, voice recognition, translation, and other tasks. Such a perspective caused an increased interest in expert architecture engineering within Deep Learning. Fueled by this interest, neural architecture search originated as a promising way to automate the tedious process of constructing a deep neural network by hand. Over the last five years, we have seen an increasing number of works focusing all efforts on studying the impact of automating deep neural network design. The spotlight has recently turned from automatically discovering classification models to other more complex tasks. Motivated by a desire for high-resolution images in real-world user-centered and expert computer vision applications, architecture search for super-resolution image restoration centers in approaches capable of automatically finding efficient and well-performing models. Here, we present a survey that, beyond delving into an overview of modern approaches to automatic neural network design, focuses on the recollection and study of neural architecture search approaches that have directed their efforts at the super-resolution image restoration tasks and future lines of research found within this emerging area of study.

如今,许多人将复杂而昂贵的神经架构视为提高现有模型在图像识别、语音识别、翻译和其他任务中性能的一种方法。这种观点使人们对深度学习中的专家架构工程越来越感兴趣。在这种兴趣的推动下,神经架构搜索应运而生,成为将手工构建深度神经网络的繁琐过程自动化的一种可行方法。在过去的五年中,我们看到越来越多的研究工作都集中在研究深度神经网络设计自动化的影响上。最近,焦点从自动发现分类模型转向了其他更复杂的任务。在现实世界中,以用户为中心的计算机视觉应用和专家计算机视觉应用对高分辨率图像的渴求推动了对超分辨率图像修复的架构探索,其核心是能够自动发现高效、性能良好的模型的方法。在此,我们将对自动神经网络设计的现代方法进行概述,并重点回顾和研究针对超分辨率图像复原任务的神经架构搜索方法,以及这一新兴研究领域的未来研究方向。
{"title":"Neural architecture search for image super-resolution: A review on the emerging state-of-the-art","authors":"","doi":"10.1016/j.neucom.2024.128481","DOIUrl":"10.1016/j.neucom.2024.128481","url":null,"abstract":"<div><p>Nowadays, complex and expensive neural architectures are seen by many as a way to improve the performance of existing models in image recognition, voice recognition, translation, and other tasks. Such a perspective caused an increased interest in expert architecture engineering within Deep Learning. Fueled by this interest, neural architecture search originated as a promising way to automate the tedious process of constructing a deep neural network by hand. Over the last five years, we have seen an increasing number of works focusing all efforts on studying the impact of automating deep neural network design. The spotlight has recently turned from automatically discovering classification models to other more complex tasks. Motivated by a desire for high-resolution images in real-world user-centered and expert computer vision applications, architecture search for super-resolution image restoration centers in approaches capable of automatically finding efficient and well-performing models. Here, we present a survey that, beyond delving into an overview of modern approaches to automatic neural network design, focuses on the recollection and study of neural architecture search approaches that have directed their efforts at the super-resolution image restoration tasks and future lines of research found within this emerging area of study.</p></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":null,"pages":null},"PeriodicalIF":5.5,"publicationDate":"2024-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142149107","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dual-loss nonlinear independent component estimation for augmenting explainable vibration samples of rotating machinery faults 用于增强旋转机械故障可解释振动样本的双损耗非线性独立分量估计
IF 5.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-31 DOI: 10.1016/j.neucom.2024.128508

Obtaining fault samples for diagnosing faults in rotating machinery for engineering applications can be costly. To address this challenge, fault sample augmentation methods are used to train fault diagnosis models. However, existing techniques mainly focus on comparing augmented signal loss with real ones, overlooking the underlying vibration mechanism of rotating machinery faults. Addressing this limitation, a novel approach, called dual-loss nonlinear independent component estimation (DLNICE), is proposed to enhance understanding of fault features in vibration signals. This integrates augmentation losses in both time and frequency domains to enrich fault vibration samples. DLNICE effectively utilizes limited fault samples for augmentation by estimating nonlinear independent components, capturing key fault characteristics like impulsiveness and cyclo-stationarity. Therefore, augmented fault samples become more explainable for analyzing rotating machinery faults. Experimental evaluations on bearing and gearbox vibration samples confirm the effectiveness of DLNICE. Utilizing the augmented samples leads to an average accuracy of 86.27 % in bearing fault diagnosis, and that of 81.60 % in gearbox fault diagnosis. The results demonstrate that DLNICE excels in augmenting high-quality vibration samples of rotating machinery faults.

在工程应用中,获取用于诊断旋转机械故障的故障样本可能成本高昂。为了应对这一挑战,人们使用故障样本增强方法来训练故障诊断模型。然而,现有技术主要侧重于比较增强信号损耗与真实信号损耗,忽略了旋转机械故障的基本振动机制。针对这一局限性,我们提出了一种名为双损耗非线性独立分量估计(DLNICE)的新方法,以增强对振动信号中故障特征的理解。这种方法同时整合了时域和频域的增强损耗,以丰富故障振动样本。DLNICE 通过估算非线性独立分量,有效地利用有限的故障样本进行增强,从而捕捉到关键的故障特征,如脉冲性和周期稳定性。因此,增强后的故障样本在分析旋转机械故障时更具可解释性。对轴承和齿轮箱振动样本的实验评估证实了 DLNICE 的有效性。利用增强样本,轴承故障诊断的平均准确率为 86.27%,齿轮箱故障诊断的平均准确率为 81.60%。结果表明,DLNICE 在增强旋转机械故障的高质量振动样本方面表现出色。
{"title":"Dual-loss nonlinear independent component estimation for augmenting explainable vibration samples of rotating machinery faults","authors":"","doi":"10.1016/j.neucom.2024.128508","DOIUrl":"10.1016/j.neucom.2024.128508","url":null,"abstract":"<div><p>Obtaining fault samples for diagnosing faults in rotating machinery for engineering applications can be costly. To address this challenge, fault sample augmentation methods are used to train fault diagnosis models. However, existing techniques mainly focus on comparing augmented signal loss with real ones, overlooking the underlying vibration mechanism of rotating machinery faults. Addressing this limitation, a novel approach, called dual-loss nonlinear independent component estimation (DLNICE), is proposed to enhance understanding of fault features in vibration signals. This integrates augmentation losses in both time and frequency domains to enrich fault vibration samples. DLNICE effectively utilizes limited fault samples for augmentation by estimating nonlinear independent components, capturing key fault characteristics like impulsiveness and cyclo-stationarity. Therefore, augmented fault samples become more explainable for analyzing rotating machinery faults. Experimental evaluations on bearing and gearbox vibration samples confirm the effectiveness of DLNICE. Utilizing the augmented samples leads to an average accuracy of 86.27 % in bearing fault diagnosis, and that of 81.60 % in gearbox fault diagnosis. The results demonstrate that DLNICE excels in augmenting high-quality vibration samples of rotating machinery faults.</p></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":null,"pages":null},"PeriodicalIF":5.5,"publicationDate":"2024-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142230660","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving AI-assisted video editing: Optimized footage analysis through multi-task learning 改进人工智能辅助视频编辑:通过多任务学习优化镜头分析
IF 5.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-30 DOI: 10.1016/j.neucom.2024.128485

In recent years, AI-assisted video editing has shown promising applications. Understanding and analyzing camera language accurately is fundamental in video editing, guiding subsequent editing and production processes. However, many existing methods for camera language analysis overlook computational efficiency and deployment requirements in favor of improving classification accuracy. Consequently, they often fail to meet the demands of scenarios with limited computing power, such as mobile devices. To address this challenge, this paper proposes an efficient multi-task camera language analysis pipeline based on shared representations. This approach employs a multi-task learning architecture with hard parameter sharing, enabling different camera language classification tasks to utilize the same low-level feature extraction network, thereby implicitly learning feature representations of the footage. Subsequently, each classification sub-task independently learns the high-level semantic information corresponding to the camera language type. This method significantly reduces computational complexity and memory usage while facilitating efficient deployment on devices with limited computing power. Furthermore, to enhance performance, we introduce a dynamic task priority strategy and a conditional dataset downsampling strategy. The experimental results demonstrate that achieved a comprehensive accuracy surpassing all previous methods. Moreover, training time was reduced by 66.33%, inference cost decreased by 59.85%, and memory usage decreased by 31.95% on the 2-task dataset MovieShots; on the 4-task dataset AVE, training time was reduced by 95.34%, inference cost decreased by 97.23%, and memory usage decreased by 61.21%.

近年来,人工智能辅助视频编辑的应用前景广阔。准确理解和分析镜头语言是视频编辑的基础,可为后续编辑和制作流程提供指导。然而,许多现有的镜头语言分析方法都忽视了计算效率和部署要求,而一味追求提高分类准确性。因此,这些方法往往无法满足计算能力有限的应用场景(如移动设备)的需求。为了应对这一挑战,本文提出了一种基于共享表征的高效多任务相机语言分析管道。该方法采用多任务学习架构,硬参数共享,使不同的摄像机语言分类任务能够利用相同的底层特征提取网络,从而隐式学习镜头的特征表征。随后,每个分类子任务独立学习与摄像机语言类型相对应的高级语义信息。这种方法大大降低了计算复杂度和内存使用量,同时有利于在计算能力有限的设备上高效部署。此外,为了提高性能,我们还引入了动态任务优先级策略和条件数据集下采样策略。实验结果表明,该方法的综合准确率超过了以往所有方法。此外,在 2 任务数据集 MovieShots 上,训练时间减少了 66.33%,推理成本减少了 59.85%,内存使用量减少了 31.95%;在 4 任务数据集 AVE 上,训练时间减少了 95.34%,推理成本减少了 97.23%,内存使用量减少了 61.21%。
{"title":"Improving AI-assisted video editing: Optimized footage analysis through multi-task learning","authors":"","doi":"10.1016/j.neucom.2024.128485","DOIUrl":"10.1016/j.neucom.2024.128485","url":null,"abstract":"<div><p>In recent years, AI-assisted video editing has shown promising applications. Understanding and analyzing camera language accurately is fundamental in video editing, guiding subsequent editing and production processes. However, many existing methods for camera language analysis overlook computational efficiency and deployment requirements in favor of improving classification accuracy. Consequently, they often fail to meet the demands of scenarios with limited computing power, such as mobile devices. To address this challenge, this paper proposes an efficient multi-task camera language analysis pipeline based on shared representations. This approach employs a multi-task learning architecture with hard parameter sharing, enabling different camera language classification tasks to utilize the same low-level feature extraction network, thereby implicitly learning feature representations of the footage. Subsequently, each classification sub-task independently learns the high-level semantic information corresponding to the camera language type. This method significantly reduces computational complexity and memory usage while facilitating efficient deployment on devices with limited computing power. Furthermore, to enhance performance, we introduce a dynamic task priority strategy and a conditional dataset downsampling strategy. The experimental results demonstrate that achieved a comprehensive accuracy surpassing all previous methods. Moreover, training time was reduced by 66.33%, inference cost decreased by 59.85%, and memory usage decreased by 31.95% on the 2-task dataset MovieShots; on the 4-task dataset AVE, training time was reduced by 95.34%, inference cost decreased by 97.23%, and memory usage decreased by 61.21%.</p></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":null,"pages":null},"PeriodicalIF":5.5,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142128682","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A review of black-box adversarial attacks on image classification 图像分类黑盒对抗攻击综述
IF 5.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-30 DOI: 10.1016/j.neucom.2024.128512

In recent years, deep learning-based image classification models have been extensively studied in academia and widely applied in industry. However, deep learning is inherently vulnerable to adversarial attacks, posing security threats to image classification models in security sensitive field, such as face recognition, medical image diagnosis and traffic sign recognition. Especially for black-box adversarial attacks, which can be carried out even without remote model information, the security issues facing deep learning are even more serious. Despite more and more attentions on this issue, existing reviews always analyze black-box adversarial attack only from one perspective, focus on only a certain application field. This paper systematically reviews and discusses existing progress, demonstrating black-box adversarial attacks from multiple perspectives and systematically classifying existing methods. Besides, we also sort out and categorize the application of current black-box adversarial attacks and identify several promising directions for future research.

近年来,基于深度学习的图像分类模型在学术界得到了广泛研究,并在工业界得到了广泛应用。然而,深度学习本身容易受到对抗性攻击,对人脸识别、医学图像诊断和交通标志识别等安全敏感领域的图像分类模型构成安全威胁。尤其是黑盒对抗攻击,即使没有远程模型信息也能实施,因此深度学习面临的安全问题更加严重。尽管这一问题受到越来越多的关注,但现有的综述总是只从一个角度分析黑盒对抗攻击,只关注某一应用领域。本文系统地回顾和讨论了现有进展,从多个角度论证了黑盒对抗攻击,并对现有方法进行了系统分类。此外,我们还对当前黑盒对抗攻击的应用进行了梳理和分类,并确定了几个有前景的未来研究方向。
{"title":"A review of black-box adversarial attacks on image classification","authors":"","doi":"10.1016/j.neucom.2024.128512","DOIUrl":"10.1016/j.neucom.2024.128512","url":null,"abstract":"<div><p>In recent years, deep learning-based image classification models have been extensively studied in academia and widely applied in industry. However, deep learning is inherently vulnerable to adversarial attacks, posing security threats to image classification models in security sensitive field, such as face recognition, medical image diagnosis and traffic sign recognition. Especially for black-box adversarial attacks, which can be carried out even without remote model information, the security issues facing deep learning are even more serious. Despite more and more attentions on this issue, existing reviews always analyze black-box adversarial attack only from one perspective, focus on only a certain application field. This paper systematically reviews and discusses existing progress, demonstrating black-box adversarial attacks from multiple perspectives and systematically classifying existing methods. Besides, we also sort out and categorize the application of current black-box adversarial attacks and identify several promising directions for future research.</p></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":null,"pages":null},"PeriodicalIF":5.5,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142149018","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unidirectional and hierarchical on-chip interconnected architecture for large-scale hardware spiking neural networks 用于大规模硬件尖峰神经网络的单向和分层片上互连架构
IF 5.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-30 DOI: 10.1016/j.neucom.2024.128480

Spiking Neural Networks (SNNs) exhibit the strong capability to address spatiotemporal dynamic problems. Recent research has explored the hardware SNN systems to solve the spatiotemporal problems in real-time. The Network-on-Chip (NoC) is an effective scheme for building large-scale hardware SNNs. However, for the existing NoC-based hardware SNNs, large area overhead and hardware power are consumed by their interconnections, because of complex topologies and router structures. Therefore, in this work a novel Unidirectional and Hierarchical on-Chip Interconnected Architecture (UHCIA) is proposed to address this problem. The proposed UHCIA mainly combines the novel hybrid topology of unidirectional multiple loops and rings, and uses a deflection router technique. Experimental results show that compared to other works, the UHCIA achieves 23.6X of area reduction and 6.4X of power reduction, with high system throughput and biological real-time computations.

尖峰神经网络(SNN)具有解决时空动态问题的强大能力。最近的研究探索了实时解决时空问题的硬件 SNN 系统。片上网络(NoC)是构建大规模硬件 SNN 的有效方案。然而,对于现有的基于 NoC 的硬件 SNN,由于复杂的拓扑结构和路由器结构,其互连会消耗大量的面积开销和硬件功耗。因此,本研究提出了一种新型单向分层片上互连架构(UHCIA)来解决这一问题。所提出的 UHCIA 主要结合了单向多环路和环路的新型混合拓扑结构,并使用了偏转路由器技术。实验结果表明,与其他研究相比,UHCIA 的面积缩小了 23.6 倍,功耗降低了 6.4 倍,系统吞吐量和生物实时计算能力都很高。
{"title":"Unidirectional and hierarchical on-chip interconnected architecture for large-scale hardware spiking neural networks","authors":"","doi":"10.1016/j.neucom.2024.128480","DOIUrl":"10.1016/j.neucom.2024.128480","url":null,"abstract":"<div><p>Spiking Neural Networks (SNNs) exhibit the strong capability to address spatiotemporal dynamic problems. Recent research has explored the hardware SNN systems to solve the spatiotemporal problems in real-time. The Network-on-Chip (NoC) is an effective scheme for building large-scale hardware SNNs. However, for the existing NoC-based hardware SNNs, large area overhead and hardware power are consumed by their interconnections, because of complex topologies and router structures. Therefore, in this work a novel Unidirectional and Hierarchical on-Chip Interconnected Architecture (UHCIA) is proposed to address this problem. The proposed UHCIA mainly combines the novel hybrid topology of unidirectional multiple loops and rings, and uses a deflection router technique. Experimental results show that compared to other works, the UHCIA achieves <span><math><mo>∼</mo></math></span>23.6X of area reduction and <span><math><mo>∼</mo></math></span>6.4X of power reduction, with high system throughput and biological real-time computations.</p></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":null,"pages":null},"PeriodicalIF":5.5,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142128530","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A comprehensive review of deep neural network interpretation using topological data analysis 利用拓扑数据分析对深度神经网络解释的全面回顾
IF 5.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-30 DOI: 10.1016/j.neucom.2024.128513

Deep neural networks have achieved significant success across various fields, but their intrinsic black-box nature hinders the further development. Addressing the interpretability challenges, topological data analysis has emerged as a promising tool to reveal these complex models. In this work, we present a review of the emerging field of interpreting deep neural networks using topological data analysis. We organize the existing body of work into distinct analytical categories, highlighting interpretations based on the topology of data, network structural characteristics, network functional characteristics, and techniques derived from Mapper. The objective of this paper is to extract the research pattern of this area, and point out the future research direction.

深度神经网络在各个领域都取得了巨大成功,但其内在的黑箱性质阻碍了其进一步发展。为应对可解释性挑战,拓扑数据分析已成为揭示这些复杂模型的一种有前途的工具。在这项工作中,我们对利用拓扑数据分析解释深度神经网络这一新兴领域进行了综述。我们将现有的工作分为不同的分析类别,重点介绍基于数据拓扑、网络结构特征、网络功能特征的解释,以及从 Mapper 衍生的技术。本文旨在提炼该领域的研究模式,并指出未来的研究方向。
{"title":"A comprehensive review of deep neural network interpretation using topological data analysis","authors":"","doi":"10.1016/j.neucom.2024.128513","DOIUrl":"10.1016/j.neucom.2024.128513","url":null,"abstract":"<div><p>Deep neural networks have achieved significant success across various fields, but their intrinsic black-box nature hinders the further development. Addressing the interpretability challenges, topological data analysis has emerged as a promising tool to reveal these complex models. In this work, we present a review of the emerging field of interpreting deep neural networks using topological data analysis. We organize the existing body of work into distinct analytical categories, highlighting interpretations based on the topology of data, network structural characteristics, network functional characteristics, and techniques derived from Mapper. The objective of this paper is to extract the research pattern of this area, and point out the future research direction.</p></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":null,"pages":null},"PeriodicalIF":5.5,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142136963","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Rainforest: A three-stage distribution adaptation framework for unsupervised time series domain adaptation 雨林:用于无监督时间序列域适应的三阶段分布适应框架
IF 5.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-30 DOI: 10.1016/j.neucom.2024.128507

Solving the unsupervised domain adaptation (UDA) task in time series is of great significance for practical applications, such as human activity recognition and machine fault diagnosis. Compared to UDA for computer vision, UDA in time series is more challenging due to the dynamics of time series data and the complex dependencies among different time steps. Existing UDA methods for time series fail to adequately capture the temporal dependencies, limiting their ability to learn domain-invariant temporal patterns. Furthermore, most UDA methods only focus on distribution adaptation on the backbone network without considering how the classifier adapts to the data distribution of the target domain. In this paper, we propose Rainforest, a three-stage UDA framework for time series. We first pre-train the backbone network through a self-supervised method called bidirectional autoregression, so that the model can comprehensively learn the temporal dependencies in time series. Next, we propose a novel meta-learning-based distribution adaptation method to achieve the joint alignment of the global and local distributions while encouraging the model to adaptively reduce the temporal dynamic differences among different domains. Finally, we design a pseudo-label-guided fine-tuning strategy to help the classifier estimate the data distribution of the target domain more accurately. Extensive experiments on four real-world time series datasets show that our Rainforest outperforms state-of-the-art methods, with an average improvement of 2.19% in accuracy and 2.41% in MF1-score.

解决时间序列中的无监督域自适应(UDA)任务对于人类活动识别和机器故障诊断等实际应用具有重要意义。与计算机视觉领域的无监督域自适应相比,时间序列领域的无监督域自适应更具挑战性,因为时间序列数据是动态的,不同时间步之间存在复杂的依赖关系。现有的时间序列 UDA 方法无法充分捕捉时间依赖性,从而限制了其学习领域不变时间模式的能力。此外,大多数 UDA 方法只关注骨干网络的分布适应,而不考虑分类器如何适应目标领域的数据分布。在本文中,我们提出了针对时间序列的三阶段 UDA 框架 Rainforest。首先,我们通过一种称为双向自回归的自监督方法对骨干网络进行预训练,从而使模型能够全面学习时间序列中的时间依赖关系。接着,我们提出了一种新颖的基于元学习的分布适应方法,以实现全局分布和局部分布的联合配准,同时鼓励模型自适应地减少不同域之间的时间动态差异。最后,我们设计了一种伪标签引导的微调策略,帮助分类器更准确地估计目标域的数据分布。在四个真实世界时间序列数据集上的广泛实验表明,我们的 Rainforest 优于最先进的方法,平均准确率提高了 2.19%,MF1 分数提高了 2.41%。
{"title":"Rainforest: A three-stage distribution adaptation framework for unsupervised time series domain adaptation","authors":"","doi":"10.1016/j.neucom.2024.128507","DOIUrl":"10.1016/j.neucom.2024.128507","url":null,"abstract":"<div><p>Solving the unsupervised domain adaptation (UDA) task in time series is of great significance for practical applications, such as human activity recognition and machine fault diagnosis. Compared to UDA for computer vision, UDA in time series is more challenging due to the dynamics of time series data and the complex dependencies among different time steps. Existing UDA methods for time series fail to adequately capture the temporal dependencies, limiting their ability to learn domain-invariant temporal patterns. Furthermore, most UDA methods only focus on distribution adaptation on the backbone network without considering how the classifier adapts to the data distribution of the target domain. In this paper, we propose Rainforest, a three-stage UDA framework for time series. We first pre-train the backbone network through a self-supervised method called bidirectional autoregression, so that the model can comprehensively learn the temporal dependencies in time series. Next, we propose a novel meta-learning-based distribution adaptation method to achieve the joint alignment of the global and local distributions while encouraging the model to adaptively reduce the temporal dynamic differences among different domains. Finally, we design a pseudo-label-guided fine-tuning strategy to help the classifier estimate the data distribution of the target domain more accurately. Extensive experiments on four real-world time series datasets show that our Rainforest outperforms state-of-the-art methods, with an average improvement of 2.19% in accuracy and 2.41% in MF1-score.</p></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":null,"pages":null},"PeriodicalIF":5.5,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142128527","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GAP: A group-based automatic pruning algorithm via convolution kernel fusion GAP:通过卷积核融合实现的基于组的自动剪枝算法
IF 5.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-30 DOI: 10.1016/j.neucom.2024.128488

In recent years, the deployment and operation of convolution neural networks on edge devices with limited computing capabilities have become increasingly challenging due to the large network structure and computational cost. Currently, the mainstream structured pruning algorithms mainly compress the network at the filter or layer level. However, these methods introduce too much human intervention with large granularities, which may lead to unpredictable performance after compression. In this paper, we propose a group-based automatic pruning algorithm(GAP) via kernel fusion to automatically search for the optimal pruning structure in a more fine-grained manner. Specifically, we first adopt a novel nonlinear dimensionality reduction clustering algorithm to divide the filters of each convolution layer into groups of equal size. Afterwards, we encode the mutual distribution similarity of the kernels within each group, and its KL divergence is employed as an importance indicator to determine the retained kernel groups through weighted fusion. Subsequently, we introduce an intelligent searching module that automatically explore and optimize the pruned structure of each layer. Finally, the pruned filters are permutated to form a dense group convolution and fine-tuned. Sufficient experiments show that, on two image classification datasets, for five advanced CNN models, our GAP algorithm outperforms most extant SOTA schemes, reduces artificial intervention, and enables efficient end-to-end training of compact models.

近年来,由于网络结构庞大、计算成本高昂,在计算能力有限的边缘设备上部署和运行卷积神经网络变得越来越具有挑战性。目前,主流的结构剪枝算法主要是在滤波器或层级别上压缩网络。然而,这些方法引入了过多的人工干预,粒度过大,可能导致压缩后的性能难以预测。在本文中,我们通过内核融合提出了一种基于组的自动剪枝算法(GAP),以更精细的方式自动搜索最佳剪枝结构。具体来说,我们首先采用一种新颖的非线性降维聚类算法,将每个卷积层的滤波器分成大小相等的组。然后,我们对每个组内内核的相互分布相似性进行编码,并将其 KL 发散作为重要性指标,通过加权融合确定保留的内核组。随后,我们引入一个智能搜索模块,自动探索和优化每一层的剪枝结构。最后,对剪枝后的滤波器进行排列组合,形成密集组卷积并进行微调。充分的实验表明,在两个图像分类数据集上,对于五种高级 CNN 模型,我们的 GAP 算法优于大多数现有的 SOTA 方案,减少了人工干预,并实现了紧凑模型的高效端到端训练。
{"title":"GAP: A group-based automatic pruning algorithm via convolution kernel fusion","authors":"","doi":"10.1016/j.neucom.2024.128488","DOIUrl":"10.1016/j.neucom.2024.128488","url":null,"abstract":"<div><p>In recent years, the deployment and operation of convolution neural networks on edge devices with limited computing capabilities have become increasingly challenging due to the large network structure and computational cost. Currently, the mainstream structured pruning algorithms mainly compress the network at the filter or layer level. However, these methods introduce too much human intervention with large granularities, which may lead to unpredictable performance after compression. In this paper, we propose a group-based automatic pruning algorithm(GAP) via kernel fusion to automatically search for the optimal pruning structure in a more fine-grained manner. Specifically, we first adopt a novel nonlinear dimensionality reduction clustering algorithm to divide the filters of each convolution layer into groups of equal size. Afterwards, we encode the mutual distribution similarity of the kernels within each group, and its KL divergence is employed as an importance indicator to determine the retained kernel groups through weighted fusion. Subsequently, we introduce an intelligent searching module that automatically explore and optimize the pruned structure of each layer. Finally, the pruned filters are permutated to form a dense group convolution and fine-tuned. Sufficient experiments show that, on two image classification datasets, for five advanced CNN models, our GAP algorithm outperforms most extant SOTA schemes, reduces artificial intervention, and enables efficient end-to-end training of compact models.</p></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":null,"pages":null},"PeriodicalIF":5.5,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142128525","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A fast strategy-solving method for adversarial team games utilizing warm starting 利用热启动的对抗性团队游戏快速策略解决方法
IF 5.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-30 DOI: 10.1016/j.neucom.2024.128509

Adversarial team games (ATGs) have garnered significant attention in recent years, leading to the emergence of various solutions such as linear programming algorithms, multi-agent reinforcement learning, and game tree transformation. The ATGs involve large-scale game trees, resulting in higher time costs for computation. In this paper, we focus on expediting the solution of team-maxmin equilibrium with correlation (TMECor), which can be considered the equilibrium that maximizes the team’s payoff. To address this, we propose a transformation with seed strategies (TSS). TSS leverages reinforcement learning to calculate player strategies. We initialize the strategies of all players, referred to as seed strategies, and incorporate them into the multi-agent game tree during the transformation process. These seed strategies serve as the starting strategies for counterfactual regret minimization (CFR). CFR initializes the strategies and cumulative regret of all players based on the seed strategy. By warm starting the whole process, our method accelerates the solving of TMECor. We conducted nine experiments using Kuhn poker and Leduc Hold’em poker. The results demonstrated that TSS improved the solving speed of TMECor.

近年来,对抗性团队博弈(ATGs)备受关注,出现了线性编程算法、多代理强化学习和博弈树转换等多种解决方案。ATG 涉及大规模博弈树,导致计算时间成本较高。在本文中,我们重点研究如何加快求解具有相关性的团队最大最小均衡(TMECor),该均衡可视为团队报酬最大化的均衡。为此,我们提出了一种种子策略转换(TSS)。TSS 利用强化学习来计算玩家策略。我们将所有玩家的策略初始化,称为种子策略,并在转换过程中将其纳入多代理博弈树。这些种子策略是反事实遗憾最小化(CFR)的起始策略。CFR 基于种子策略初始化所有博弈者的策略和累积遗憾。通过温暖启动整个过程,我们的方法加快了 TMECor 的求解速度。 我们使用库恩扑克和勒杜克扑克进行了九次实验。结果表明,TSS 提高了 TMECor 的求解速度。
{"title":"A fast strategy-solving method for adversarial team games utilizing warm starting","authors":"","doi":"10.1016/j.neucom.2024.128509","DOIUrl":"10.1016/j.neucom.2024.128509","url":null,"abstract":"<div><p>Adversarial team games (ATGs) have garnered significant attention in recent years, leading to the emergence of various solutions such as linear programming algorithms, multi-agent reinforcement learning, and game tree transformation. The ATGs involve large-scale game trees, resulting in higher time costs for computation. In this paper, we focus on expediting the solution of team-maxmin equilibrium with correlation (TMECor), which can be considered the equilibrium that maximizes the team’s payoff. To address this, we propose a transformation with seed strategies (TSS). TSS leverages reinforcement learning to calculate player strategies. We initialize the strategies of all players, referred to as seed strategies, and incorporate them into the multi-agent game tree during the transformation process. These seed strategies serve as the starting strategies for counterfactual regret minimization (CFR). CFR initializes the strategies and cumulative regret of all players based on the seed strategy. By warm starting the whole process, our method accelerates the solving of TMECor. We conducted nine experiments using Kuhn poker and Leduc Hold’em poker. The results demonstrated that TSS improved the solving speed of TMECor.</p></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":null,"pages":null},"PeriodicalIF":5.5,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142149020","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Neurocomputing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1