首页 > 最新文献

Artificial Intelligence Review最新文献

英文 中文
Cloud-edge-end collaborative caching and UAV-assisted offloading decision based on the fusion of deep reinforcement learning algorithms 基于融合深度强化学习算法的云边缘协同缓存和无人机辅助卸载决策
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-05 DOI: 10.1007/s10462-025-11391-8
Sifeng Zhu, Zhaowei Song, Changlong Huang, Rui Qiao, Hai Zhu

The cloud-edge-end collaboration system provides a new impetus for the development of intelligent transportation. In order to optimize the quality of service for intelligent transportation system users and improve system resource utilization, A three-tier caching strategy for cloud-edge-end collaboration based on efficiency collaboration task popularity (CSEPCA) was proposed, which exploits server resource characteristics and performs fine-grained cache replacements based on real-time task popularity to address the challenges associated with balancing server cache space and cost. To achieve an optimal balance between server cache space and cost, the problem of determining the availability of server cache space is formulated as a constrained markov decision process (CMDP), and an enhanced deep reinforcement learning algorithm based on soft updating (AT-SAC) was designed to achieve multi-objective optimization of system latency, energy consumption, and resource depletion rate, with the aim of improving service response speed and enhancing user service quality. To address challenges in effectively serving vehicles in areas with weak communication signals from cloud-edge servers, UAV swarms were introduced to assist with vehicle task offloading computations. A comprehensive optimization algorithm (Co-DRL-P) was proposed, which integrates enhanced deep reinforcement Learning (ERDDPG) and improved particle swarm optimization (A-PSO) algorithms to optimize UAV trajectories and communication angles, aiming to deliver superior service quality to users. Finally, we evaluate the performance of the proposed scheme through comprehensive simulation experiments. Specifically, when the number of users is 30, the system latency of the proposed scheme is 17.9%, 11.5%, 2.6%, and 60.2% lower than baseline schemes such as DQN, DDPG, TD3, and collaborative randomized schemes, and the system energy consumption is reduced by 20.6%, 15.9%, 9.4%, and 129.9%. Notably, the overall system cost for drone-assisted user offloading is reduced by approximately 49.6% in areas with weak cloud server signals.

云-端协同系统为智能交通发展提供了新的动力。为了优化智能交通系统用户的服务质量,提高系统资源利用率,提出了一种基于效率协同任务流行度的云边缘协同三层缓存策略,该策略利用服务器资源特性,基于实时任务流行度进行细粒度缓存替换,解决了服务器缓存空间与成本平衡的难题。为实现服务器缓存空间与成本之间的最优平衡,将服务器缓存空间可用性确定问题归结为约束马尔可夫决策过程(CMDP),设计了基于软更新的增强型深度强化学习算法(AT-SAC),实现了系统延迟、能耗和资源消耗率的多目标优化,以提高服务响应速度和用户服务质量。为了解决在云边缘服务器通信信号较弱的地区为车辆提供有效服务的挑战,引入了无人机群来辅助车辆任务卸载计算。提出了一种综合优化算法(Co-DRL-P),该算法集成了增强深度强化学习(ERDDPG)和改进粒子群优化(A- pso)算法,对无人机轨迹和通信角度进行优化,旨在为用户提供卓越的服务质量。最后,通过综合仿真实验对所提方案的性能进行了评价。其中,当用户数为30时,所提方案的系统时延比DQN、DDPG、TD3和协同随机化等基准方案分别降低17.9%、11.5%、2.6%和60.2%,系统能耗分别降低20.6%、15.9%、9.4%和129.9%。值得注意的是,在云服务器信号较弱的地区,无人机辅助用户卸载的整体系统成本降低了约49.6%。
{"title":"Cloud-edge-end collaborative caching and UAV-assisted offloading decision based on the fusion of deep reinforcement learning algorithms","authors":"Sifeng Zhu,&nbsp;Zhaowei Song,&nbsp;Changlong Huang,&nbsp;Rui Qiao,&nbsp;Hai Zhu","doi":"10.1007/s10462-025-11391-8","DOIUrl":"10.1007/s10462-025-11391-8","url":null,"abstract":"<div><p>The cloud-edge-end collaboration system provides a new impetus for the development of intelligent transportation. In order to optimize the quality of service for intelligent transportation system users and improve system resource utilization, A three-tier caching strategy for cloud-edge-end collaboration based on efficiency collaboration task popularity (CSEPCA) was proposed, which exploits server resource characteristics and performs fine-grained cache replacements based on real-time task popularity to address the challenges associated with balancing server cache space and cost. To achieve an optimal balance between server cache space and cost, the problem of determining the availability of server cache space is formulated as a constrained markov decision process (CMDP), and an enhanced deep reinforcement learning algorithm based on soft updating (AT-SAC) was designed to achieve multi-objective optimization of system latency, energy consumption, and resource depletion rate, with the aim of improving service response speed and enhancing user service quality. To address challenges in effectively serving vehicles in areas with weak communication signals from cloud-edge servers, UAV swarms were introduced to assist with vehicle task offloading computations. A comprehensive optimization algorithm (Co-DRL-P) was proposed, which integrates enhanced deep reinforcement Learning (ERDDPG) and improved particle swarm optimization (A-PSO) algorithms to optimize UAV trajectories and communication angles, aiming to deliver superior service quality to users. Finally, we evaluate the performance of the proposed scheme through comprehensive simulation experiments. Specifically, when the number of users is 30, the system latency of the proposed scheme is 17.9%, 11.5%, 2.6%, and 60.2% lower than baseline schemes such as DQN, DDPG, TD3, and collaborative randomized schemes, and the system energy consumption is reduced by 20.6%, 15.9%, 9.4%, and 129.9%. Notably, the overall system cost for drone-assisted user offloading is reduced by approximately 49.6% in areas with weak cloud server signals.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"58 12","pages":""},"PeriodicalIF":13.9,"publicationDate":"2025-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11391-8.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145456681","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Scientific approach to problem solving-inspired optimization of stacking ensemble learning for enhanced civil engineering informatics 基于问题求解的科学方法——基于叠加集成学习的优化土木工程信息学
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-04 DOI: 10.1007/s10462-025-11356-x
Dinh-Nhat Truong, Jui-Sheng Chou

This study introduces the Scientific Approach to Problem Solving-inspired Optimization (SAPSO) algorithm, a novel metaheuristic specifically designed for applications in civil engineering informatics. SAPSO imitates the structured process of scientific inquiry—covering problem review, hypothesis formulation, data collection, and analysis—to systematically explore complex search spaces. This approach enables SAPSO to reliably identify global optima. The algorithm’s performance was extensively tested against eleven leading metaheuristic algorithms using the IEEE Congress on Evolutionary Computation benchmark suites from 2020 (CEC 2020) and 2022 (CEC 2022). The comparison included the Artificial Bee Colony, Cultural Algorithm, Genetic Algorithm, Differential Evolution, Artificial Gorilla Troops Optimizer, Grey Wolf Optimizer, Particle Swarm Optimization, Red Kite Optimization Algorithm, Symbiotic Organisms Search, Teaching–Learning-Based Optimization, and Whale Optimization Algorithm. Statistical analysis with the Wilcoxon rank-sum test confirmed SAPSO’s superior results across these benchmarks. Additionally, this study presents a stacked ensemble machine learning framework called the SAPSO-Weighted Features Stacking System (SAPSO-WFSS), which combines SAPSO with two predictive models: a Radial Basis Function Neural Network and Least Squares Support Vector Regression. SAPSO is used to optimize both feature weights and model hyperparameters. Experiments on five diverse civil engineering case studies show that SAPSO-WFSS provides high accuracy, with Mean Absolute Percentage Error values as low as 2.4%, outperforming traditional methods. These findings demonstrate SAPSO’s potential as a powerful tool for improving prediction reliability in infrastructure maintenance and solving complex optimization problems in civil engineering.

本研究介绍了问题解决启发优化(SAPSO)算法的科学方法,这是一种专门为土木工程信息学应用而设计的新型元启发式算法。SAPSO模仿科学调查的结构化过程,包括问题审查,假设制定,数据收集和分析,以系统地探索复杂的搜索空间。这种方法使SAPSO能够可靠地识别全局最优。利用2020年(CEC 2020)和2022年(CEC 2022)的IEEE进化计算大会基准套件,针对11种领先的元启发式算法对该算法的性能进行了广泛测试。其中包括人工蜂群算法、文化算法、遗传算法、差分进化算法、人工大猩猩优化算法、灰狼优化算法、粒子群优化算法、红风筝优化算法、共生生物搜索算法、基于教与学的优化算法、鲸鱼优化算法。使用Wilcoxon秩和检验的统计分析证实了SAPSO在这些基准测试中的优越结果。此外,本研究提出了一种称为SAPSO加权特征堆叠系统(SAPSO- wfss)的堆叠集成机器学习框架,该框架将SAPSO与两种预测模型(径向基函数神经网络和最小二乘支持向量回归)相结合。SAPSO用于优化特征权值和模型超参数。5个不同土木工程案例的实验表明,SAPSO-WFSS具有较高的精度,平均绝对百分比误差值低至2.4%,优于传统方法。这些发现证明了SAPSO作为提高基础设施维护预测可靠性和解决土木工程中复杂优化问题的强大工具的潜力。
{"title":"Scientific approach to problem solving-inspired optimization of stacking ensemble learning for enhanced civil engineering informatics","authors":"Dinh-Nhat Truong,&nbsp;Jui-Sheng Chou","doi":"10.1007/s10462-025-11356-x","DOIUrl":"10.1007/s10462-025-11356-x","url":null,"abstract":"<div><p>This study introduces the Scientific Approach to Problem Solving-inspired Optimization (SAPSO) algorithm, a novel metaheuristic specifically designed for applications in civil engineering informatics. SAPSO imitates the structured process of scientific inquiry—covering problem review, hypothesis formulation, data collection, and analysis—to systematically explore complex search spaces. This approach enables SAPSO to reliably identify global optima. The algorithm’s performance was extensively tested against eleven leading metaheuristic algorithms using the IEEE Congress on Evolutionary Computation benchmark suites from 2020 (CEC 2020) and 2022 (CEC 2022). The comparison included the Artificial Bee Colony, Cultural Algorithm, Genetic Algorithm, Differential Evolution, Artificial Gorilla Troops Optimizer, Grey Wolf Optimizer, Particle Swarm Optimization, Red Kite Optimization Algorithm, Symbiotic Organisms Search, Teaching–Learning-Based Optimization, and Whale Optimization Algorithm. Statistical analysis with the Wilcoxon rank-sum test confirmed SAPSO’s superior results across these benchmarks. Additionally, this study presents a stacked ensemble machine learning framework called the SAPSO-Weighted Features Stacking System (SAPSO-WFSS), which combines SAPSO with two predictive models: a Radial Basis Function Neural Network and Least Squares Support Vector Regression. SAPSO is used to optimize both feature weights and model hyperparameters. Experiments on five diverse civil engineering case studies show that SAPSO-WFSS provides high accuracy, with Mean Absolute Percentage Error values as low as 2.4%, outperforming traditional methods. These findings demonstrate SAPSO’s potential as a powerful tool for improving prediction reliability in infrastructure maintenance and solving complex optimization problems in civil engineering.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"58 12","pages":""},"PeriodicalIF":13.9,"publicationDate":"2025-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11356-x.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145456177","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cyberswarm: a novel swarm intelligence algorithm inspired by cyber community dynamics 网络群体:一种受网络社区动态启发的新型群体智能算法
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-04 DOI: 10.1007/s10462-025-11406-4
Abdelsadeq Elfergany, Ammar Adl, Mohammed Kayed

Recommendation systems face challenges in dynamically adapting to evolving user preferences and interactions within complex social networks. Traditional approaches often fail to account for the intricate interactions within cyber-social systems and lack the flexibility to generalize across diverse domains, highlighting the need for more adaptive and versatile solutions. In this work, we introduce a general-purpose swarm intelligence algorithm for recommendation systems, designed to adapt seamlessly to varying applications. It was inspired by social psychology principles. The framework models user preferences and community influences within a dynamic hypergraph structure. It leverages centrality-based feature extraction and Node2Vec embeddings. Preference evolution is guided by message-passing mechanisms and hierarchical graph modeling, enabling real-time adaptation to changing behaviors. Experimental evaluations demonstrated the algorithm’s superior performance in various recommendation tasks, including social networks and content discovery. Key metrics such as Hit Rate (HR), Mean Reciprocal Rank (MRR), and Normalized Discounted Cumulative Gain (NDCG) consistently outperformed baseline methods across multiple datasets. The model’s adaptability to dynamic environments allowed for contextually relevant and precise recommendations. The proposed algorithm represents an advancement in recommendation systems by bridging individual preferences and community influences. Its general-purpose design enables applications in diverse domains, including social graphs, personalized learning, and medical graphs. This work highlights the potential of integrating swarm intelligence with network dynamics to address complex optimization challenges in recommendation systems.

在复杂的社会网络中,推荐系统面临着动态适应不断变化的用户偏好和交互的挑战。传统的方法往往不能解释网络社会系统内部复杂的相互作用,并且缺乏在不同领域推广的灵活性,这突出了对更具适应性和通用性的解决方案的需求。在这项工作中,我们为推荐系统引入了一种通用的群体智能算法,旨在无缝地适应不同的应用。它的灵感来自社会心理学原理。该框架在动态超图结构中对用户偏好和社区影响进行建模。它利用了基于中心性的特征提取和Node2Vec嵌入。偏好演化由消息传递机制和分层图建模指导,支持实时适应不断变化的行为。实验评估表明,该算法在各种推荐任务中表现优异,包括社交网络和内容发现。关键指标,如命中率(HR)、平均互惠等级(MRR)和标准化贴现累积增益(NDCG)在多个数据集上始终优于基线方法。该模型对动态环境的适应性允许提供与上下文相关且精确的建议。该算法通过连接个人偏好和社区影响,代表了推荐系统的一种进步。它的通用设计支持各种领域的应用程序,包括社交图、个性化学习和医学图。这项工作强调了将群体智能与网络动力学相结合的潜力,以解决推荐系统中复杂的优化挑战。
{"title":"Cyberswarm: a novel swarm intelligence algorithm inspired by cyber community dynamics","authors":"Abdelsadeq Elfergany,&nbsp;Ammar Adl,&nbsp;Mohammed Kayed","doi":"10.1007/s10462-025-11406-4","DOIUrl":"10.1007/s10462-025-11406-4","url":null,"abstract":"<p>Recommendation systems face challenges in dynamically adapting to evolving user preferences and interactions within complex social networks. Traditional approaches often fail to account for the intricate interactions within cyber-social systems and lack the flexibility to generalize across diverse domains, highlighting the need for more adaptive and versatile solutions. In this work, we introduce a general-purpose swarm intelligence algorithm for recommendation systems, designed to adapt seamlessly to varying applications. It was inspired by social psychology principles. The framework models user preferences and community influences within a dynamic hypergraph structure. It leverages centrality-based feature extraction and Node2Vec embeddings. Preference evolution is guided by message-passing mechanisms and hierarchical graph modeling, enabling real-time adaptation to changing behaviors. Experimental evaluations demonstrated the algorithm’s superior performance in various recommendation tasks, including social networks and content discovery. Key metrics such as Hit Rate (HR), Mean Reciprocal Rank (MRR), and Normalized Discounted Cumulative Gain (NDCG) consistently outperformed baseline methods across multiple datasets. The model’s adaptability to dynamic environments allowed for contextually relevant and precise recommendations. The proposed algorithm represents an advancement in recommendation systems by bridging individual preferences and community influences. Its general-purpose design enables applications in diverse domains, including social graphs, personalized learning, and medical graphs. This work highlights the potential of integrating swarm intelligence with network dynamics to address complex optimization challenges in recommendation systems.</p>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"58 12","pages":""},"PeriodicalIF":13.9,"publicationDate":"2025-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11406-4.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145456175","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Philosophical proposition optimizer (ΦPO): an epistemology-inspired algorithm for numerical optimization 哲学命题优化器(ΦPO):一个受认识论启发的数值优化算法
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-04 DOI: 10.1007/s10462-025-11383-8
Siamak Talatahari, Hadi Bayazidi, Pooya Sareh

The development of intelligent optimization methods has become a highly active research area in recent decades. This paper introduces a philosophy-inspired optimization algorithm called the Philosophical Proposition Optimizer (ΦPO), which models knowledge acquisition based on philosophical propositions in epistemology. In the proposed philosophical model, three developmental states for philosophical propositions, Justified True Belief (JTB), Possibly False Belief (PFB), and Unjustified True Belief (UTB), are iteratively refined using three specialized operators: Providing Justification (PJ), Raising Metaphysical Skepticism (RMS), and Raising Epistemic Skepticism (RES). To evaluate the performance of ΦPO on challenging optimization problems, it is applied to the single-objective bound-constrained benchmark problems of the IEEE Congress on Evolutionary Computation 2014 and 2024 (CEC 2014 and 2024), as well as to benchmark engineering problems. The performance of ΦPO is compared against five categories of algorithms: (1) widely used classical methods, (2) established post-2019 methods, (3) advanced PSO- and DE-based methods, (4) winners of CEC competitions, and (5) well-studied methods for solving engineering design problems. Two established non-parametric statistical methods, the Friedman test and the Wilcoxon signed-rank test, are used to analyze performance. The findings highlight the advantages of ΦPO across a range of numerical optimization problems, underscoring its competitiveness and potential in the field. Importantly, ΦPO was intentionally designed to be simple, interpretable, and parameter-free, avoiding complex adaptive strategies and extensive parameter tuning. It consistently delivers stable, high-quality solutions and exhibits fast convergence in many cases. The results demonstrate that ΦPO performs competitively across multiple benchmark suites, often ranking among the top-performing algorithms and outperforming several state-of-the-art methods, including recent CEC competition winners and engineering-specific optimizers. Its unique epistemic approach to solution refinement further enhances robustness, distinguishing it in both numerical and engineering optimization tasks.

近几十年来,智能优化方法的发展成为一个非常活跃的研究领域。本文介绍了一种哲学启发的优化算法——哲学命题优化器(ΦPO),它基于认识论中的哲学命题来建模知识获取。在提出的哲学模型中,哲学命题的三种发展状态,即被证明的真信念(JTB),可能的假信念(PFB)和未被证明的真信念(UTB),通过三个专门的操作符:提供证明(PJ),提出形而上学怀疑(RMS)和提出认知怀疑(RES)来迭代改进。为了评估ΦPO在具有挑战性的优化问题上的性能,将其应用于IEEE进化计算大会2014和2024 (CEC 2014和2024)的单目标约束基准问题,以及基准工程问题。将ΦPO的性能与五类算法进行比较:(1)广泛使用的经典方法,(2)2019年后建立的方法,(3)先进的基于PSO和de的方法,(4)CEC竞赛的获奖者,以及(5)研究充分的解决工程设计问题的方法。使用两种已建立的非参数统计方法,Friedman检验和Wilcoxon有符号秩检验来分析性能。研究结果突出了ΦPO在一系列数值优化问题上的优势,强调了其在该领域的竞争力和潜力。重要的是,ΦPO被有意设计成简单、可解释和无参数的,避免了复杂的自适应策略和广泛的参数调优。它始终如一地提供稳定、高质量的解决方案,并在许多情况下表现出快速收敛。结果表明,ΦPO在多个基准测试套件中具有竞争力,通常在性能最佳的算法中名列前茅,并且优于几种最先进的方法,包括最近的CEC竞赛获胜者和工程特定的优化器。其独特的求解方法进一步增强了鲁棒性,使其在数值和工程优化任务中脱颖而出。
{"title":"Philosophical proposition optimizer (ΦPO): an epistemology-inspired algorithm for numerical optimization","authors":"Siamak Talatahari,&nbsp;Hadi Bayazidi,&nbsp;Pooya Sareh","doi":"10.1007/s10462-025-11383-8","DOIUrl":"10.1007/s10462-025-11383-8","url":null,"abstract":"<div><p>The development of intelligent optimization methods has become a highly active research area in recent decades. This paper introduces a philosophy-inspired optimization algorithm called the Philosophical Proposition Optimizer (ΦPO), which models knowledge acquisition based on philosophical propositions in epistemology. In the proposed philosophical model, three developmental states for philosophical propositions, Justified True Belief (JTB), Possibly False Belief (PFB), and Unjustified True Belief (UTB), are iteratively refined using three specialized operators: Providing Justification (PJ), Raising Metaphysical Skepticism (RMS), and Raising Epistemic Skepticism (RES). To evaluate the performance of ΦPO on challenging optimization problems, it is applied to the single-objective bound-constrained benchmark problems of the IEEE Congress on Evolutionary Computation 2014 and 2024 (CEC 2014 and 2024), as well as to benchmark engineering problems. The performance of ΦPO is compared against five categories of algorithms: (1) widely used classical methods, (2) established post-2019 methods, (3) advanced PSO- and DE-based methods, (4) winners of CEC competitions, and (5) well-studied methods for solving engineering design problems. Two established non-parametric statistical methods, the Friedman test and the Wilcoxon signed-rank test, are used to analyze performance. The findings highlight the advantages of ΦPO across a range of numerical optimization problems, underscoring its competitiveness and potential in the field. Importantly, ΦPO was intentionally designed to be simple, interpretable, and parameter-free, avoiding complex adaptive strategies and extensive parameter tuning. It consistently delivers stable, high-quality solutions and exhibits fast convergence in many cases. The results demonstrate that ΦPO performs competitively across multiple benchmark suites, often ranking among the top-performing algorithms and outperforming several state-of-the-art methods, including recent CEC competition winners and engineering-specific optimizers. Its unique epistemic approach to solution refinement further enhances robustness, distinguishing it in both numerical and engineering optimization tasks.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"58 12","pages":""},"PeriodicalIF":13.9,"publicationDate":"2025-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11383-8.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145456176","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Writing style change detection: state of the art, challenges, and research opportunities 写作风格变化检测:现状、挑战和研究机会
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-31 DOI: 10.1007/s10462-025-11402-8
Abeer Saad Alsheddi, Mohamed El Bachir Menai

A style change detection (SCD) task finds the locations of writing style changes within multi-authored documents. In 2017, the SCD task was introduced to assist in various applications, including cybercrime and literary analysis. This paper examines several document representation and prediction methods, including statistical, deep neural network (DNN), classical machine learning (ML), and hybrid methods. An analysis of existing datasets and the performance of SCD solutions is also included in this review. The findings demonstrate that the best method for attaining high performance is supervised ML, especially feed-forward neural networks with pretrained-based representations. Even though DNN models work well for other tasks, they are less frequently developed for this task, and their results need to be improved. This study details a set of challenges that are related to selecting features and adopting pretrained models. Additionally, the available literature is fairly limited in this task compared to others. For further investigation, several research directions are highlighted, including the design of SCD datasets with more realistic styles and the exploration of various learning techniques. This work encourages researchers to foster the growth and development of this research area.

样式更改检测(SCD)任务查找多作者文档中书写样式更改的位置。2017年,SCD任务被引入,以协助各种应用,包括网络犯罪和文学分析。本文研究了几种文档表示和预测方法,包括统计、深度神经网络(DNN)、经典机器学习(ML)和混合方法。本综述还包括对现有数据集和SCD解决方案性能的分析。研究结果表明,获得高性能的最佳方法是有监督的机器学习,特别是具有基于预训练的表征的前馈神经网络。尽管DNN模型在其他任务中工作得很好,但它们很少用于这项任务,并且它们的结果需要改进。本研究详细介绍了与选择特征和采用预训练模型相关的一系列挑战。此外,与其他任务相比,在此任务中可用的文献相当有限。在进一步的研究中,强调了几个研究方向,包括设计更逼真风格的SCD数据集和探索各种学习技术。这项工作鼓励研究人员促进该研究领域的成长和发展。
{"title":"Writing style change detection: state of the art, challenges, and research opportunities","authors":"Abeer Saad Alsheddi,&nbsp;Mohamed El Bachir Menai","doi":"10.1007/s10462-025-11402-8","DOIUrl":"10.1007/s10462-025-11402-8","url":null,"abstract":"<div><p>A style change detection (SCD) task finds the locations of writing style changes within multi-authored documents. In 2017, the SCD task was introduced to assist in various applications, including cybercrime and literary analysis. This paper examines several document representation and prediction methods, including statistical, deep neural network (DNN), classical machine learning (ML), and hybrid methods. An analysis of existing datasets and the performance of SCD solutions is also included in this review. The findings demonstrate that the best method for attaining high performance is supervised ML, especially feed-forward neural networks with pretrained-based representations. Even though DNN models work well for other tasks, they are less frequently developed for this task, and their results need to be improved. This study details a set of challenges that are related to selecting features and adopting pretrained models. Additionally, the available literature is fairly limited in this task compared to others. For further investigation, several research directions are highlighted, including the design of SCD datasets with more realistic styles and the exploration of various learning techniques. This work encourages researchers to foster the growth and development of this research area.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"58 12","pages":""},"PeriodicalIF":13.9,"publicationDate":"2025-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11402-8.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145406230","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A review of artificial intelligence in herbarium specimen image analysis 人工智能在植物标本图像分析中的研究进展
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-31 DOI: 10.1007/s10462-025-11408-2
Yu-Yue Guo, Haibin Cai, Gemma L. C. Bramley, Hannah J. Atkins, Baihua Li, Stephanos Theodossiades

The digitisation of hundreds of millions of herbarium specimen images and their labels has created an unprecedented resource for taxonomy, ecology, and conservation, motivating the development of artificial intelligence (AI) solutions. Automated analysis of these high-resolution scans faces significant challenges, including data imbalance, information loss, model interpretability and explainability, and scalable Open-Set Recognition (OSR). This paper provides an in-depth algorithm-level review of AI methodologies for herbarium image classification, tracing the development from classical classification models like Convolutional Neural Networks (CNNs) and Vision Transformers (ViTs) to cutting-edge multimodal frameworks. In addition to classification, the review further investigates vision-based analytical tasks critical to herbarium image analysis, including specimen image segmentation, label text identification using Large Language Models (LLMs), and Human-in-the-Loop (HITL) quality assurance strategies. Furthermore, this review reveals practical challenges in specimen image analysis along with their promising solutions and potential future directions.

数以亿计的植物标本馆标本图像及其标签的数字化为分类学、生态学和保护创造了前所未有的资源,推动了人工智能(AI)解决方案的发展。这些高分辨率扫描的自动化分析面临着重大挑战,包括数据不平衡、信息丢失、模型可解释性和可解释性以及可扩展的开放集识别(OSR)。本文对植物标本馆图像分类的人工智能方法进行了深入的算法级回顾,追溯了从卷积神经网络(cnn)和视觉变形器(ViTs)等经典分类模型到前沿多模态框架的发展。除了分类之外,该综述还进一步研究了基于视觉的分析任务,包括标本图像分割,使用大型语言模型(LLMs)的标签文本识别,以及人在环(HITL)质量保证策略。此外,这篇综述揭示了标本图像分析的实际挑战,以及它们有希望的解决方案和潜在的未来方向。
{"title":"A review of artificial intelligence in herbarium specimen image analysis","authors":"Yu-Yue Guo,&nbsp;Haibin Cai,&nbsp;Gemma L. C. Bramley,&nbsp;Hannah J. Atkins,&nbsp;Baihua Li,&nbsp;Stephanos Theodossiades","doi":"10.1007/s10462-025-11408-2","DOIUrl":"10.1007/s10462-025-11408-2","url":null,"abstract":"<div><p>The digitisation of hundreds of millions of herbarium specimen images and their labels has created an unprecedented resource for taxonomy, ecology, and conservation, motivating the development of artificial intelligence (AI) solutions. Automated analysis of these high-resolution scans faces significant challenges, including data imbalance, information loss, model interpretability and explainability, and scalable Open-Set Recognition (OSR). This paper provides an in-depth algorithm-level review of AI methodologies for herbarium image classification, tracing the development from classical classification models like Convolutional Neural Networks (CNNs) and Vision Transformers (ViTs) to cutting-edge multimodal frameworks. In addition to classification, the review further investigates vision-based analytical tasks critical to herbarium image analysis, including specimen image segmentation, label text identification using Large Language Models (LLMs), and Human-in-the-Loop (HITL) quality assurance strategies. Furthermore, this review reveals practical challenges in specimen image analysis along with their promising solutions and potential future directions.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"58 12","pages":""},"PeriodicalIF":13.9,"publicationDate":"2025-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11408-2.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145406229","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Datasets for large language models: a comprehensive survey 大型语言模型的数据集:综合调查
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-31 DOI: 10.1007/s10462-025-11403-7
Yang Liu, Jiahuan Cao, Chongyu Liu, Kai Ding, Lianwen Jin

This paper embarks on an exploration into the large language model (LLM) datasets, which play a crucial role in the remarkable advancements of LLMs. The datasets serve as the foundational infrastructure analogous to a root system that sustains and nurtures the development of LLMs. Consequently, examination of these datasets emerges as a critical topic in research. In order to address the current lack of a comprehensive overview and thorough analysis of LLM datasets, and to gain insights into their current status and future trends, this survey consolidates and categorizes the fundamental aspects of LLM datasets from four perspectives: (a) pre-training corpora; (b) instruction fine-tuning datasets; (c) preference datasets; (d) evaluation datasets. The survey sheds light on the prevailing challenges and points out potential avenues for future investigation. Additionally, a comprehensive review of the existing available dataset resources is also provided, including statistics from 303 datasets, covering 8 language categories and spanning 32 domains. Information from 20 dimensions is incorporated into the dataset statistics. The total data size surveyed surpasses 774.5 TB for pre-training corpora and 700 M instances for other datasets. We aim to present the entire landscape of LLM text datasets, serving as a comprehensive reference for researchers in this field and contributing to future studies. Related resources are available at: https://github.com/lmmlzn/Awesome-LLMs-Datasets.

本文对大型语言模型(LLM)数据集进行了探讨,这些数据集在LLM的显著进步中起着至关重要的作用。数据集作为基础设施,类似于维持和培育法学硕士发展的根系统。因此,对这些数据集的检查成为研究中的一个关键主题。为了解决目前缺乏对法学硕士数据集的全面概述和深入分析的问题,并深入了解法学硕士数据集的现状和未来趋势,本调查从四个方面对法学硕士数据集的基本方面进行了整合和分类:(a)预训练语料库;(b)指令微调数据集;(c)偏好数据集;(d)评估数据集。该调查揭示了当前的挑战,并指出了未来调查的潜在途径。此外,还提供了对现有可用数据集资源的全面审查,包括来自303个数据集的统计数据,涵盖8个语言类别和跨越32个领域。来自20个维度的信息被合并到数据集统计中。调查的总数据大小超过了预训练语料库的774.5 TB和其他数据集的700 M实例。我们的目标是呈现法学硕士文本数据集的整个景观,为该领域的研究人员提供全面的参考,并为未来的研究做出贡献。相关资源可从https://github.com/lmmlzn/Awesome-LLMs-Datasets获得。
{"title":"Datasets for large language models: a comprehensive survey","authors":"Yang Liu,&nbsp;Jiahuan Cao,&nbsp;Chongyu Liu,&nbsp;Kai Ding,&nbsp;Lianwen Jin","doi":"10.1007/s10462-025-11403-7","DOIUrl":"10.1007/s10462-025-11403-7","url":null,"abstract":"<div><p>This paper embarks on an exploration into the large language model (LLM) datasets, which play a crucial role in the remarkable advancements of LLMs. The datasets serve as the foundational infrastructure analogous to a root system that sustains and nurtures the development of LLMs. Consequently, examination of these datasets emerges as a critical topic in research. In order to address the current lack of a comprehensive overview and thorough analysis of LLM datasets, and to gain insights into their current status and future trends, this survey consolidates and categorizes the fundamental aspects of LLM datasets from four perspectives: (a) pre-training corpora; (b) instruction fine-tuning datasets; (c) preference datasets; (d) evaluation datasets. The survey sheds light on the prevailing challenges and points out potential avenues for future investigation. Additionally, a comprehensive review of the existing available dataset resources is also provided, including statistics from 303 datasets, covering 8 language categories and spanning 32 domains. Information from 20 dimensions is incorporated into the dataset statistics. The total data size surveyed surpasses 774.5 TB for pre-training corpora and 700 M instances for other datasets. We aim to present the entire landscape of LLM text datasets, serving as a comprehensive reference for researchers in this field and contributing to future studies. Related resources are available at: https://github.com/lmmlzn/Awesome-LLMs-Datasets.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"58 12","pages":""},"PeriodicalIF":13.9,"publicationDate":"2025-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11403-7.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145406419","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A survey on large language models unlearning: taxonomy, evaluations, and future directions 大型语言模型学习综述:分类、评估和未来方向
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-29 DOI: 10.1007/s10462-025-11376-7
Uyen N. Le-Khac, Vinh N. X. Truong

Following the introduction of data privacy regulations and “the right to be forgotten”, large language models (LLMs) unlearning has emerged as a promising data removal solution for compliance purposes, while also facilitating a diverse range of applications, including copyright protection, model detoxification and correction, and jailbreaking defence. In this survey, we present the taxonomy of existing LLMs unlearning algorithms, summarise unlearning evaluation methods including specialised benchmarks and threat models, and explore the applications of unlearning to provide a broad overview of the current state-of-the-art. We propose a novel problem formulation of LLMs unlearning with the additional unlearning objective: “robustness” to reflect the growing research interest in not only effectively and efficiently eliminating unwanted data, but also ensuring the process is performed safely and securely. To the best of our knowledge, we are the first to examine the robustness of unlearning algorithms as well as threat models for robustness evaluation, aspects that have not been assessed in past surveys. We also identify the limitations of the current approaches, including limited applicability to black-box models, vulnerability to adversarial attacks and knowledge leakage, and inefficiency, all of which require further improvement in future works. Furthermore, our survey highlights future directions for LLMs unlearning research, such as the development of comprehensive evaluation benchmarks, the movement towards robust unlearning and explainable AI for unlearning mechanisms, and addressing potential ethical dilemmas in unlearning governance.

在引入数据隐私法规和“被遗忘权”之后,大型语言模型(llm)遗忘已成为一种有前途的数据删除解决方案,用于合规目的,同时也促进了各种应用,包括版权保护,模型解毒和纠正以及越狱防御。在本调查中,我们提出了现有法学硕士学习算法的分类,总结了学习评估方法,包括专门的基准和威胁模型,并探索了学习的应用,以提供当前最新技术的广泛概述。我们提出了一种新的llm学习问题公式,附带了额外的学习目标:“鲁棒性”,以反映对不仅有效和高效地消除不需要的数据,而且确保过程安全可靠地执行的日益增长的研究兴趣。据我们所知,我们是第一个研究解除学习算法的鲁棒性以及鲁棒性评估的威胁模型的,这些方面在过去的调查中没有被评估过。我们还指出了当前方法的局限性,包括对黑盒模型的有限适用性,对抗性攻击和知识泄漏的脆弱性,以及效率低下,所有这些都需要在未来的工作中进一步改进。此外,我们的调查还强调了法学硕士学习研究的未来方向,例如综合评估基准的发展,向强大的学习和可解释的人工智能的发展,以及解决学习治理中潜在的伦理困境。
{"title":"A survey on large language models unlearning: taxonomy, evaluations, and future directions","authors":"Uyen N. Le-Khac,&nbsp;Vinh N. X. Truong","doi":"10.1007/s10462-025-11376-7","DOIUrl":"10.1007/s10462-025-11376-7","url":null,"abstract":"<div><p>Following the introduction of data privacy regulations and “the right to be forgotten”, large language models (LLMs) unlearning has emerged as a promising data removal solution for compliance purposes, while also facilitating a diverse range of applications, including copyright protection, model detoxification and correction, and jailbreaking defence. In this survey, we present the taxonomy of existing LLMs unlearning algorithms, summarise unlearning evaluation methods including specialised benchmarks and threat models, and explore the applications of unlearning to provide a broad overview of the current state-of-the-art. We propose a novel problem formulation of LLMs unlearning with the additional unlearning objective: “robustness” to reflect the growing research interest in not only effectively and efficiently eliminating unwanted data, but also ensuring the process is performed safely and securely. To the best of our knowledge, we are the first to examine the robustness of unlearning algorithms as well as threat models for robustness evaluation, aspects that have not been assessed in past surveys. We also identify the limitations of the current approaches, including limited applicability to black-box models, vulnerability to adversarial attacks and knowledge leakage, and inefficiency, all of which require further improvement in future works. Furthermore, our survey highlights future directions for LLMs unlearning research, such as the development of comprehensive evaluation benchmarks, the movement towards robust unlearning and explainable AI for unlearning mechanisms, and addressing potential ethical dilemmas in unlearning governance.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"58 12","pages":""},"PeriodicalIF":13.9,"publicationDate":"2025-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11376-7.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145406144","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep learning for multivariate time series anomaly detection: an evaluation of reconstruction-based methods 多元时间序列异常检测的深度学习:基于重建方法的评价
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-29 DOI: 10.1007/s10462-025-11401-9
Mohammed A. Yahya, Antonio R. Moya, Sebastián Ventura

In the field of anomaly detection in time series, remarkable advances based on deep learning methodologies and, more specifically, reconstruction-based methods have been proposed. These methods are particularly valuable, as they can capture the fundamental structure of the data and enable the detection of subtle anomalies that traditional techniques might overlook. Reviews of the existing literature discuss anomaly detection from a general perspective and consider a specific anomaly type, hindering the process of proposing new and better algorithms. This paper focuses on reconstruction-based methods in isolation, as they have been demonstrated to present the best performance of the three main groups in deep anomaly detection models described so far. Thus, it intends to extend the literature by addressing in detail reconstruction-based methods for anomaly detection in multivariate time series, to provide richer information about these methods, and to include extensive experimentation, not usually performed in existing surveys on the topic. Finally, the paper presents useful insights (strengths and weaknesses of the methods) extracted from the experimental study, and significant challenges and future research directions related to the potential of these methods in anomaly detection.

在时间序列异常检测领域,基于深度学习方法,更具体地说,基于重建的方法已经取得了显著进展。这些方法特别有价值,因为它们可以捕获数据的基本结构,并能够检测到传统技术可能忽略的细微异常。现有文献综述从一般角度讨论异常检测,并考虑特定的异常类型,阻碍了提出新的和更好的算法的过程。本文的重点是基于重建的方法,因为它们已经被证明是迄今为止描述的深度异常检测模型中三种主要方法的最佳性能。因此,本文打算通过详细介绍多变量时间序列中基于重构的异常检测方法来扩展文献,提供有关这些方法的更丰富的信息,并包括在该主题的现有调查中通常不进行的广泛实验。最后,本文提出了从实验研究中提取的有用见解(方法的优点和缺点),以及与这些方法在异常检测中的潜力相关的重大挑战和未来的研究方向。
{"title":"Deep learning for multivariate time series anomaly detection: an evaluation of reconstruction-based methods","authors":"Mohammed A. Yahya,&nbsp;Antonio R. Moya,&nbsp;Sebastián Ventura","doi":"10.1007/s10462-025-11401-9","DOIUrl":"10.1007/s10462-025-11401-9","url":null,"abstract":"<div><p>In the field of anomaly detection in time series, remarkable advances based on deep learning methodologies and, more specifically, reconstruction-based methods have been proposed. These methods are particularly valuable, as they can capture the fundamental structure of the data and enable the detection of subtle anomalies that traditional techniques might overlook. Reviews of the existing literature discuss anomaly detection from a general perspective and consider a specific anomaly type, hindering the process of proposing new and better algorithms. This paper focuses on reconstruction-based methods in isolation, as they have been demonstrated to present the best performance of the three main groups in deep anomaly detection models described so far. Thus, it intends to extend the literature by addressing in detail reconstruction-based methods for anomaly detection in multivariate time series, to provide richer information about these methods, and to include extensive experimentation, not usually performed in existing surveys on the topic. Finally, the paper presents useful insights (strengths and weaknesses of the methods) extracted from the experimental study, and significant challenges and future research directions related to the potential of these methods in anomaly detection.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"58 12","pages":""},"PeriodicalIF":13.9,"publicationDate":"2025-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11401-9.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145406328","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-sequence discrete recursive descent optimizer: a novel optimization algorithm based on quasi-Newton directions and imprecise search step lengths 多序列离散递归下降优化器:一种基于准牛顿方向和不精确搜索步长的新型优化算法
IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-28 DOI: 10.1007/s10462-025-11384-7
Hanqiao Huang, Bo Du, Bo Zhang, Huan Zhou, Gang Hu

An original mathematical theory-inspired optimization algorithm, named multi-sequence discrete recursive descent optimizer (MDRDO) is proposed for numerical optimization and engineering issues. The main inspiration is derived from the deterministic recursive descent method, composed of descent directions and step lengths. Firstly, by adopting multiple initial sequences, a multiple-sequence searching framework (MSF) is constructed to explore the feasible space fully. Then, according to the Taylor expansion, the approximate discrete gradient is calculated to form a discrete quasi-Newton direction (BFGS) and the imprecise search step length for each sequence. Furthermore, to enhance the ability to escape local optimums, two random searching strategies are employed to keep the diversity of candidate solutions. Hence, the MDRDO is constructed by merging the quasi-Newton direction, imprecise step length and random strategies, in which suitable descent directions and step lengths help the MDRDO to explore a promising region and exploit the optimal solutions. The effectiveness of the MDRDO is verified by the comparison experiments with other 10 excellent algorithms on two typical test suites. Results reveal that the MDRDO outperforms others for 62.07% among 29 functions of CEC2017 test suite and 80% among 10 functions of CEC2020 test suite. Average convergence curves prove the ability of MDRDO to approach better solutions quickly. The applicability of the MDRDO is tested on 5 engineering problems with the other 9 widely used algorithms, in which the MDRDO ranks first for all problems with excellent comprehensive performance. Overall, the test results demonstrate that MDRDO is a promising tool for addressing practical optimization problems.

针对数值优化和工程问题,提出了一种新颖的基于数学理论的多序列离散递归下降优化算法(MDRDO)。主要灵感来源于确定性递归下降法,由下降方向和步长组成。首先,采用多个初始序列,构造多序列搜索框架(MSF),充分探索可行空间;然后,根据泰勒展开,计算近似离散梯度,形成离散的准牛顿方向(BFGS)和每个序列的不精确搜索步长。此外,为了提高逃避局部最优的能力,采用了两种随机搜索策略来保持候选解的多样性。因此,融合准牛顿方向、不精确步长和随机策略构建了MDRDO,其中合适的下降方向和步长有助于MDRDO探索有希望的区域并获得最优解。在两个典型的测试套件上与其他10种优秀算法进行了对比实验,验证了MDRDO算法的有效性。结果表明,在CEC2017测试套件的29个功能中,MDRDO优于其他测试套件的62.07%,在CEC2020测试套件的10个功能中,MDRDO优于其他测试套件的80%。平均收敛曲线证明了MDRDO快速逼近较优解的能力。与其他9种广泛使用的算法对5个工程问题进行了适用性测试,其中MDRDO以优异的综合性能在所有问题中排名第一。总体而言,测试结果表明MDRDO是解决实际优化问题的一个有前途的工具。
{"title":"Multi-sequence discrete recursive descent optimizer: a novel optimization algorithm based on quasi-Newton directions and imprecise search step lengths","authors":"Hanqiao Huang,&nbsp;Bo Du,&nbsp;Bo Zhang,&nbsp;Huan Zhou,&nbsp;Gang Hu","doi":"10.1007/s10462-025-11384-7","DOIUrl":"10.1007/s10462-025-11384-7","url":null,"abstract":"<div><p>An original mathematical theory-inspired optimization algorithm, named multi-sequence discrete recursive descent optimizer (MDRDO) is proposed for numerical optimization and engineering issues. The main inspiration is derived from the deterministic recursive descent method, composed of descent directions and step lengths. Firstly, by adopting multiple initial sequences, a multiple-sequence searching framework (MSF) is constructed to explore the feasible space fully. Then, according to the Taylor expansion, the approximate discrete gradient is calculated to form a discrete quasi-Newton direction (BFGS) and the imprecise search step length for each sequence. Furthermore, to enhance the ability to escape local optimums, two random searching strategies are employed to keep the diversity of candidate solutions. Hence, the MDRDO is constructed by merging the quasi-Newton direction, imprecise step length and random strategies, in which suitable descent directions and step lengths help the MDRDO to explore a promising region and exploit the optimal solutions. The effectiveness of the MDRDO is verified by the comparison experiments with other 10 excellent algorithms on two typical test suites. Results reveal that the MDRDO outperforms others for 62.07% among 29 functions of CEC2017 test suite and 80% among 10 functions of CEC2020 test suite. Average convergence curves prove the ability of MDRDO to approach better solutions quickly. The applicability of the MDRDO is tested on 5 engineering problems with the other 9 widely used algorithms, in which the MDRDO ranks first for all problems with excellent comprehensive performance. Overall, the test results demonstrate that MDRDO is a promising tool for addressing practical optimization problems.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"58 12","pages":""},"PeriodicalIF":13.9,"publicationDate":"2025-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11384-7.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145406312","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Artificial Intelligence Review
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1