首页 > 最新文献

Intelligent Systems with Applications最新文献

英文 中文
Enhancing lung disease diagnosis with deep-learning-based CT scan image segmentation 基于深度学习的CT扫描图像分割增强肺部疾病诊断
IF 4.3 Pub Date : 2025-09-01 Epub Date: 2025-07-28 DOI: 10.1016/j.iswa.2025.200565
Rima Tri Wahyuningrum , Achmad Bauravindah , Indah Agustien Siradjuddin , Budi Dwi Satoto , Amillia Kartika Sari , Anggraini Dwi Sensusiati
The coronavirus disease 2019 (COVID-19) pandemic has underscored the need for efficient diagnostic methods owing to the limitations in sensitivity and time constraints associated with molecular tests such as reverse transcription PCR (RT-PCR). This research aims to enhance the efficiency of COVID-19 and other lung diseases such as pneumonia, tuberculosis, bronchitis, emphysema, asthma, and others diagnoses. As an alternative diagnostic, we considered an approach based on enhanced computed tomography (CT) scan images using deep learning (DL). However, we propose a preprocessing segmentation method to enhance the accuracy of DL-based classification that uses the UNet++ architecture, an encoder-decoder approach in DL. In this architecture, the encoder reduces the image resolution to extract informative feature maps while the decoder returns the resolution to the original size. UNet++ is available in four levels: UNet++ L1, L2, L3, and L4, and its performance is compared to that of several other models, including SegNet, FCANet, and DeepLabV3+. Using two different datasets, RSPHC (Indonesia) and Kaggle, testing was conducted to determine the model with the optimum performance. The criteria used to evaluate model performance included the Dice coefficient and IoU metrics, most efficient computational time, and minimal resource requirements (measured by trainable parameters). The UNet++ L4 model achieved a Dice coefficient of 0.994, IoU of 0.989, computational time of 0.925 s, and 9.16 million trainable parameters on the RSPHC dataset. Whereas on the Kaggle dataset it achieved a Dice coefficient of 0.961, IoU of 0.930, computational time of 1.189 s, and 9.16 million trainable parameters. Therefore, the UNet++ L4 model is ideal for accurate segmentation, computational efficiency, and affordable resource requirements. Thus, this research improves lung disease diagnosis through enhanced CT scan images using DL.
由于反转录PCR (RT-PCR)等分子检测在灵敏度和时间上的限制,2019年冠状病毒病(COVID-19)大流行凸显了高效诊断方法的必要性。本研究旨在提高COVID-19以及肺炎、肺结核、支气管炎、肺气肿、哮喘等其他肺部疾病的诊断效率。作为一种替代诊断,我们考虑了一种基于增强计算机断层扫描(CT)扫描图像的方法,该方法使用深度学习(DL)。然而,我们提出了一种预处理分割方法来提高基于DL的分类的准确性,该方法使用UNet++架构,这是DL中的一种编码器-解码器方法。在这种架构中,编码器降低图像分辨率以提取信息特征映射,而解码器将分辨率返回到原始大小。unnet++有4个级别:unnet++ L1、L2、L3和L4,其性能与其他几种模型(包括SegNet、FCANet和DeepLabV3+)进行了比较。利用RSPHC(印度尼西亚)和Kaggle两个不同的数据集进行测试,以确定具有最佳性能的模型。用于评估模型性能的标准包括Dice系数和IoU指标、最有效的计算时间和最小的资源需求(通过可训练参数测量)。unet++ L4模型在RSPHC数据集上的Dice系数为0.994,IoU为0.989,计算时间为0.925 s,可训练参数为916万个。而在Kaggle数据集上,它的Dice系数为0.961,IoU为0.930,计算时间为1189 s,可训练参数为916万个。因此,UNet++ L4模型是精确分割、计算效率和可负担的资源需求的理想选择。因此,本研究通过DL增强CT扫描图像提高肺部疾病的诊断。
{"title":"Enhancing lung disease diagnosis with deep-learning-based CT scan image segmentation","authors":"Rima Tri Wahyuningrum ,&nbsp;Achmad Bauravindah ,&nbsp;Indah Agustien Siradjuddin ,&nbsp;Budi Dwi Satoto ,&nbsp;Amillia Kartika Sari ,&nbsp;Anggraini Dwi Sensusiati","doi":"10.1016/j.iswa.2025.200565","DOIUrl":"10.1016/j.iswa.2025.200565","url":null,"abstract":"<div><div>The coronavirus disease 2019 (COVID-19) pandemic has underscored the need for efficient diagnostic methods owing to the limitations in sensitivity and time constraints associated with molecular tests such as reverse transcription PCR (RT-PCR). This research aims to enhance the efficiency of COVID-19 and other lung diseases such as pneumonia, tuberculosis, bronchitis, emphysema, asthma, and others diagnoses. As an alternative diagnostic, we considered an approach based on enhanced computed tomography (CT) scan images using deep learning (DL). However, we propose a preprocessing segmentation method to enhance the accuracy of DL-based classification that uses the UNet++ architecture, an encoder-decoder approach in DL. In this architecture, the encoder reduces the image resolution to extract informative feature maps while the decoder returns the resolution to the original size. UNet++ is available in four levels: UNet++ L1, L2, L3, and L4, and its performance is compared to that of several other models, including SegNet, FCANet, and DeepLabV3+. Using two different datasets, RSPHC (Indonesia) and Kaggle, testing was conducted to determine the model with the optimum performance. The criteria used to evaluate model performance included the Dice coefficient and IoU metrics, most efficient computational time, and minimal resource requirements (measured by trainable parameters). The UNet++ L4 model achieved a Dice coefficient of 0.994, IoU of 0.989, computational time of 0.925 s, and 9.16 million trainable parameters on the RSPHC dataset. Whereas on the Kaggle dataset it achieved a Dice coefficient of 0.961, IoU of 0.930, computational time of 1.189 s, and 9.16 million trainable parameters. Therefore, the UNet++ L4 model is ideal for accurate segmentation, computational efficiency, and affordable resource requirements. Thus, this research improves lung disease diagnosis through enhanced CT scan images using DL.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"27 ","pages":"Article 200565"},"PeriodicalIF":4.3,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144766984","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Role and Applications of Semantic Interoperability Tools and eXplainable AI in the Development of Smart Food Systems: Findings from a Systematic Literature Review 语义互操作性工具和可解释的人工智能在智能食品系统发展中的作用和应用:来自系统文献综述的发现
Pub Date : 2025-09-01 Epub Date: 2025-06-27 DOI: 10.1016/j.iswa.2025.200547
Donika Xhani, Gayane Sedrakyan, Anand Gavai, Renata Guizzardi, Jos van Hillegersberg
Smart food systems generate vast and diverse data across the supply chain, yet inconsistent data structures and limited interoperability hinder their full potential. Achieving semantic interoperability, where systems can exchange and interpret data with shared meaning, is essential for enabling intelligent integration and decision-making. Tools such as ontologies, knowledge graphs, and reasoning engines play a key role in this process. In this paper, we refer to these as Semantic Interoperability (SI) tools: a broad category that includes technologies grounded in Semantic Web standards (e.g., RDF, OWL, SPARQL) but emphasizes their applied role in aligning meaning across heterogeneous systems. Coupled with eXplainable Artificial Intelligence (XAI), these technologies enhance transparency and trust in AI-driven decisions, such as personalized food recommendations tailored to an individual’s health conditions and preferences. This paper presents a Systematic Literature Review (SLR) examining the role of semantic interoperability tools and XAI in the development of smart food systems. Through an analysis of 39 studies, the review identifies key semantic technologies and XAI methods used in food systems, with a focus on their application in intelligent food recommendation systems. The findings reveal that while significant progress has been made, current systems often lack adequate transparency and personalization, limiting user trust and engagement. To address these gaps, the paper proposes the integration of semantic interoperability tools with XAI to create smarter, more reliable food systems. As part of this effort, the paper introduces the conceptual model for the Semantic Explainable Food Recommendation Ontology (SEFRO), a work-in-progress ontology, designed to connect entities and relationships within food systems in an intelligent manner, with the goal of enabling personalized, explainable, and interoperable food recommendations that meet the growing demands for smart food systems.
智能食品系统在整个供应链中产生大量不同的数据,但不一致的数据结构和有限的互操作性阻碍了它们充分发挥潜力。实现语义互操作性(系统可以交换和解释具有共享含义的数据)对于实现智能集成和决策至关重要。诸如本体、知识图和推理引擎之类的工具在这个过程中起着关键作用。在本文中,我们将这些工具称为语义互操作性(Semantic Interoperability, SI)工具:这是一个广泛的类别,包括基于语义Web标准的技术(例如,RDF、OWL、SPARQL),但强调它们在跨异构系统调整含义方面的应用作用。再加上可解释人工智能(XAI),这些技术提高了人工智能驱动决策的透明度和信任度,例如根据个人健康状况和偏好量身定制的个性化食品建议。本文介绍了一篇系统文献综述(SLR),研究了语义互操作性工具和人工智能在智能食品系统开发中的作用。通过对39项研究的分析,本文确定了食品系统中使用的关键语义技术和XAI方法,并重点介绍了它们在智能食品推荐系统中的应用。调查结果显示,虽然取得了重大进展,但目前的系统往往缺乏足够的透明度和个性化,限制了用户的信任和参与。为了解决这些差距,本文提出将语义互操作性工具与XAI集成,以创建更智能、更可靠的食品系统。作为这项工作的一部分,本文介绍了语义可解释食品推荐本体(SEFRO)的概念模型,SEFRO是一个正在开发的本体,旨在以智能的方式连接食品系统中的实体和关系,目标是实现个性化、可解释和可互操作的食品推荐,以满足对智能食品系统日益增长的需求。
{"title":"The Role and Applications of Semantic Interoperability Tools and eXplainable AI in the Development of Smart Food Systems: Findings from a Systematic Literature Review","authors":"Donika Xhani,&nbsp;Gayane Sedrakyan,&nbsp;Anand Gavai,&nbsp;Renata Guizzardi,&nbsp;Jos van Hillegersberg","doi":"10.1016/j.iswa.2025.200547","DOIUrl":"10.1016/j.iswa.2025.200547","url":null,"abstract":"<div><div>Smart food systems generate vast and diverse data across the supply chain, yet inconsistent data structures and limited interoperability hinder their full potential. Achieving semantic interoperability, where systems can exchange and interpret data with shared meaning, is essential for enabling intelligent integration and decision-making. Tools such as ontologies, knowledge graphs, and reasoning engines play a key role in this process. In this paper, we refer to these as <em>Semantic Interoperability (SI) tools</em>: a broad category that includes technologies grounded in Semantic Web standards (e.g., RDF, OWL, SPARQL) but emphasizes their applied role in aligning meaning across heterogeneous systems. Coupled with eXplainable Artificial Intelligence (XAI), these technologies enhance transparency and trust in AI-driven decisions, such as personalized food recommendations tailored to an individual’s health conditions and preferences. This paper presents a Systematic Literature Review (SLR) examining the role of semantic interoperability tools and XAI in the development of smart food systems. Through an analysis of 39 studies, the review identifies key semantic technologies and XAI methods used in food systems, with a focus on their application in intelligent food recommendation systems. The findings reveal that while significant progress has been made, current systems often lack adequate transparency and personalization, limiting user trust and engagement. To address these gaps, the paper proposes the integration of semantic interoperability tools with XAI to create smarter, more reliable food systems. As part of this effort, the paper introduces the conceptual model for the Semantic Explainable Food Recommendation Ontology (SEFRO), a work-in-progress ontology, designed to connect entities and relationships within food systems in an intelligent manner, with the goal of enabling personalized, explainable, and interoperable food recommendations that meet the growing demands for smart food systems.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"27 ","pages":"Article 200547"},"PeriodicalIF":0.0,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144548423","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Monkeypox optimizer: A TinyML bio-inspired evolutionary optimization algorithm and its engineering applications 猴痘优化器:TinyML生物启发的进化优化算法及其工程应用
Pub Date : 2025-09-01 Epub Date: 2025-07-15 DOI: 10.1016/j.iswa.2025.200557
Marwa F. Mohamed , Ahmed Hamed
High-dimensional optimization remains a key challenge in computational intelligence, especially under resource constraints. Evolutionary algorithms, which mimic the change in heritable characteristics of biological populations, have been proposed to address this. These algorithms apply selection pressure to favor better solutions over generations, and stochastic variations may occasionally introduce suboptimal candidates to preserve population diversity. However, they often struggle to balance exploration and exploitation, leading to suboptimal solutions, premature convergence, and significant computational demands, making them unsuitable for resource-constrained environments. This paper introduces Monkeypox Optimization (MO), a novel evolutionary algorithm inspired by the infection and replication lifecycle of the monkeypox virus. MO mimics the virus’s rapid spread by employing virus-to-cell infection, where the virus persistently seeks out vulnerable cells to penetrate—representing global exploration of the search space. Once inside, cell-to-cell transmission enables fast local propagation, modeling the refinement of high-potential solutions through accelerated replication. To conserve resources, MO continuously deletes the least effective virion copies, maintaining a compact and memory-efficient population. This biologically grounded design not only accelerates convergence but also aligns MO with TinyML principles, making it ideally suited for low-power, resource-constrained IoT environments. MO is benchmarked against 21 recent algorithms across 90 functions from CEC-2017, CEC-2019, and CEC-2020, and validated on three engineering design problems. Results show MO achieves up to 13% lower energy consumption and 34% shorter execution time compared to state-of-the-art competitors, while maintaining robust accuracy. A theoretical analysis reveals MO’s time complexity is O(mn+RTn), confirming its scalability. Statistical validation via Friedman and Fisher tests further supports MO’s performance gains.
高维优化仍然是计算智能的一个关键挑战,特别是在资源限制下。为了解决这个问题,已经提出了模拟生物种群遗传特征变化的进化算法。这些算法施加选择压力,在几代人中倾向于更好的解决方案,随机变化可能偶尔会引入次优候选方案,以保持种群多样性。然而,它们往往难以平衡勘探和开发,导致次优解决方案、过早收敛和大量的计算需求,使它们不适合资源受限的环境。猴痘优化算法(Monkeypox Optimization, MO)是一种受猴痘病毒感染和复制生命周期启发的新型进化算法。MO通过病毒对细胞感染来模拟病毒的快速传播,病毒持续寻找易受攻击的细胞进行渗透——这代表了对搜索空间的全球探索。一旦进入细胞内部,细胞间传输可以实现快速本地传播,通过加速复制模拟高潜力解决方案的改进。为了节省资源,MO不断地删除最无效的病毒粒子副本,保持紧凑和内存高效的种群。这种基于生物的设计不仅加速了融合,而且使MO与TinyML原则保持一致,使其非常适合低功耗,资源受限的物联网环境。MO以21种最新算法为基准,涵盖cecc -2017、cecc -2019和cecc -2020的90个功能,并在三个工程设计问题上进行了验证。结果表明,与最先进的竞争对手相比,MO的能耗降低了13%,执行时间缩短了34%,同时保持了强大的准确性。理论分析表明,MO的时间复杂度为0 (mn+RTn),证实了其可扩展性。通过Friedman和Fisher测试的统计验证进一步支持MO的性能提升。
{"title":"Monkeypox optimizer: A TinyML bio-inspired evolutionary optimization algorithm and its engineering applications","authors":"Marwa F. Mohamed ,&nbsp;Ahmed Hamed","doi":"10.1016/j.iswa.2025.200557","DOIUrl":"10.1016/j.iswa.2025.200557","url":null,"abstract":"<div><div>High-dimensional optimization remains a key challenge in computational intelligence, especially under resource constraints. Evolutionary algorithms, which mimic the change in heritable characteristics of biological populations, have been proposed to address this. These algorithms apply selection pressure to favor better solutions over generations, and stochastic variations may occasionally introduce suboptimal candidates to preserve population diversity. However, they often struggle to balance exploration and exploitation, leading to suboptimal solutions, premature convergence, and significant computational demands, making them unsuitable for resource-constrained environments. This paper introduces Monkeypox Optimization (MO), a novel evolutionary algorithm inspired by the infection and replication lifecycle of the monkeypox virus. MO mimics the virus’s rapid spread by employing virus-to-cell infection, where the virus persistently seeks out vulnerable cells to penetrate—representing global exploration of the search space. Once inside, cell-to-cell transmission enables fast local propagation, modeling the refinement of high-potential solutions through accelerated replication. To conserve resources, MO continuously deletes the least effective virion copies, maintaining a compact and memory-efficient population. This biologically grounded design not only accelerates convergence but also aligns MO with TinyML principles, making it ideally suited for low-power, resource-constrained IoT environments. MO is benchmarked against 21 recent algorithms across 90 functions from CEC-2017, CEC-2019, and CEC-2020, and validated on three engineering design problems. Results show MO achieves up to 13% lower energy consumption and 34% shorter execution time compared to state-of-the-art competitors, while maintaining robust accuracy. A theoretical analysis reveals MO’s time complexity is <span><math><mrow><mi>O</mi><mrow><mo>(</mo><mi>m</mi><mi>n</mi><mo>+</mo><mi>R</mi><mi>T</mi><mi>n</mi><mo>)</mo></mrow></mrow></math></span>, confirming its scalability. Statistical validation via Friedman and Fisher tests further supports MO’s performance gains.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"27 ","pages":"Article 200557"},"PeriodicalIF":0.0,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144687060","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Advancements and challenges of deep learning architectures for aerial image analysis: A systematic review 航空图像分析中深度学习架构的进步与挑战:系统综述
Pub Date : 2025-09-01 Epub Date: 2025-06-04 DOI: 10.1016/j.iswa.2025.200537
Hashibul Ahsan Shoaib , Hadiur Rahman Nabil , Md Anisur Rahman , Md Mohsin Kabir , M.F. Mridha , Jungpil Shin
The rapid advancement of deep learning (DL) technologies has significantly transformed the domain of aerial image analysis. This systematic review explores the forefront of deep learning architectures specifically designed for the processing and analysis of aerial imagery. It offers a comprehensive examination of updated models, such as Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), Generative Adversarial Networks (GANs), and Transformers, highlighting their unique contributions and comparative effectiveness in aerial image analysis. This review critically compares these architectures through an extensive literature survey, focusing on their impact on enhancing accuracy, computational efficiency, and overall performance in critical aerial imaging tasks, such as classification, object detection, and semantic segmentation. Additionally, it sheds light on the innovative architectural improvements that have been crucial in overcoming traditional challenges associated with aerial image processing, such as handling high-resolution data, managing diverse and changing landscapes, and ensuring real-time analysis capabilities. By synthesizing current findings and identifying prevailing trends, this review not only charts the progress in the field but also outlines future research directions, emphasizing the need for more adaptable, robust, and efficient deep-learning solutions to meet the growing demands of aerial image analysis.
深度学习(DL)技术的快速发展极大地改变了航空图像分析领域。本系统综述探讨了专门为航空图像处理和分析设计的深度学习架构的前沿。它提供了一个全面的检查更新的模型,如卷积神经网络(cnn),循环神经网络(rnn),生成对抗网络(gan),和变形,突出他们在航空图像分析的独特贡献和比较有效性。这篇综述通过广泛的文献调查对这些架构进行了批判性的比较,重点关注它们对提高准确性、计算效率和关键航空成像任务(如分类、目标检测和语义分割)的整体性能的影响。此外,它还揭示了创新的建筑改进,这些改进对于克服与航空图像处理相关的传统挑战至关重要,例如处理高分辨率数据,管理多样化和不断变化的景观,以及确保实时分析能力。通过综合目前的研究结果和确定流行趋势,本文不仅概述了该领域的进展,还概述了未来的研究方向,强调需要更具适应性、鲁棒性和效率的深度学习解决方案,以满足日益增长的航空图像分析需求。
{"title":"Advancements and challenges of deep learning architectures for aerial image analysis: A systematic review","authors":"Hashibul Ahsan Shoaib ,&nbsp;Hadiur Rahman Nabil ,&nbsp;Md Anisur Rahman ,&nbsp;Md Mohsin Kabir ,&nbsp;M.F. Mridha ,&nbsp;Jungpil Shin","doi":"10.1016/j.iswa.2025.200537","DOIUrl":"10.1016/j.iswa.2025.200537","url":null,"abstract":"<div><div>The rapid advancement of deep learning (DL) technologies has significantly transformed the domain of aerial image analysis. This systematic review explores the forefront of deep learning architectures specifically designed for the processing and analysis of aerial imagery. It offers a comprehensive examination of updated models, such as Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), Generative Adversarial Networks (GANs), and Transformers, highlighting their unique contributions and comparative effectiveness in aerial image analysis. This review critically compares these architectures through an extensive literature survey, focusing on their impact on enhancing accuracy, computational efficiency, and overall performance in critical aerial imaging tasks, such as classification, object detection, and semantic segmentation. Additionally, it sheds light on the innovative architectural improvements that have been crucial in overcoming traditional challenges associated with aerial image processing, such as handling high-resolution data, managing diverse and changing landscapes, and ensuring real-time analysis capabilities. By synthesizing current findings and identifying prevailing trends, this review not only charts the progress in the field but also outlines future research directions, emphasizing the need for more adaptable, robust, and efficient deep-learning solutions to meet the growing demands of aerial image analysis.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"27 ","pages":"Article 200537"},"PeriodicalIF":0.0,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144272578","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Design and development of a dexterous soft-robotics based assistive exoglove with kinematic modeling 基于柔性机器人的辅助手套的设计与开发
Pub Date : 2025-09-01 Epub Date: 2025-07-04 DOI: 10.1016/j.iswa.2025.200550
Nawara Mahmood Broti , Shamim Ahmed Deowan , A.S.M. Shamsul Arefin
Necessity of a fully functional hand in our life is beyond description. Yet, a portion of the population is unable to move and control their hand due to paralysis. An assistive device can aid both daily activities and rehabilitation. This paper presents a dexterous soft robotics-based assistive glove with spatial kinematic model and control system. Unlike existing designs, our proposed five-fingered glove provides 20 degrees of freedom (DoFs), closely resembling a human hand. Each finger has 4 DoFs with controlled flexion, extension, abduction, and adduction motion ability. The tendon-driven mechanism simplifies design and control, while 3D-printed thermoplastic polyurethane (TPU) material ensures comfort, lightness, and an anthropomorphic appearance. The derived forward and inverse kinematics of each finger are capable of mapping joint angles to fingertip positions and orientations. To validate the kinematic model, virtual simulation was conducted to confirm its accuracy; while basic hand functionality experiments proved the gloves’ effectiveness. We expect this research to contribute to medical robotics, biomechanics, and assistive technology.
在我们的生活中,一只功能齐全的手的必要性是无法形容的。然而,由于瘫痪,一部分人无法移动和控制他们的手。辅助装置可以帮助日常活动和康复。提出了一种基于柔性机器人的灵巧辅助手套,并建立了空间运动学模型和控制系统。与现有的设计不同,我们提出的五指手套提供了20个自由度(DoFs),与人手非常相似。每根手指有4个自由度,可控制屈伸外展内收运动能力。肌腱驱动的机构简化了设计和控制,而3d打印的热塑性聚氨酯(TPU)材料确保了舒适性、轻盈性和拟人化的外观。导出的每个手指的正运动学和逆运动学能够映射关节角度到指尖的位置和方向。为验证运动学模型的正确性,对其进行了虚拟仿真;而基本的手部功能实验证明了手套的有效性。我们期望这项研究对医疗机器人、生物力学和辅助技术做出贡献。
{"title":"Design and development of a dexterous soft-robotics based assistive exoglove with kinematic modeling","authors":"Nawara Mahmood Broti ,&nbsp;Shamim Ahmed Deowan ,&nbsp;A.S.M. Shamsul Arefin","doi":"10.1016/j.iswa.2025.200550","DOIUrl":"10.1016/j.iswa.2025.200550","url":null,"abstract":"<div><div>Necessity of a fully functional hand in our life is beyond description. Yet, a portion of the population is unable to move and control their hand due to paralysis. An assistive device can aid both daily activities and rehabilitation. This paper presents a dexterous soft robotics-based assistive glove with spatial kinematic model and control system. Unlike existing designs, our proposed five-fingered glove provides 20 degrees of freedom (DoFs), closely resembling a human hand. Each finger has 4 DoFs with controlled flexion, extension, abduction, and adduction motion ability. The tendon-driven mechanism simplifies design and control, while 3D-printed thermoplastic polyurethane (TPU) material ensures comfort, lightness, and an anthropomorphic appearance. The derived forward and inverse kinematics of each finger are capable of mapping joint angles to fingertip positions and orientations. To validate the kinematic model, virtual simulation was conducted to confirm its accuracy; while basic hand functionality experiments proved the gloves’ effectiveness. We expect this research to contribute to medical robotics, biomechanics, and assistive technology.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"27 ","pages":"Article 200550"},"PeriodicalIF":0.0,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144597231","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A multi-source multi-layer-based transfer learning approach for forecasting customer demands of newly launched products 基于多源多层的新产品客户需求预测迁移学习方法
Pub Date : 2025-09-01 Epub Date: 2025-06-27 DOI: 10.1016/j.iswa.2025.200548
Supriyo Ahmed , Ripon K. Chakrabortty , Daryl L. Essam
Forecasting the future demand for newly launched products has been challenging for supply chain practitioners, often due to the lack of data. However, market surveys and extracting knowledge by examining similar market products to find the behaviour of a new product can be inaccurate and lead to erroneous results, which ultimately lead to a misestimation of the overall cost of a business. Meanwhile, with the advancement of artificial intelligence (AI) approaches, such as Transfer Learning (TL), this misestimation of cost can be reduced by more accurately forecasting the demand for newly launched products by seeking knowledge from the historical data of other similar products. Consequently, this paper investigates several classical AI-based TL approaches to predict customer demand for new products and stores. Thereafter, a novel Multi-Source Multi-Layer Transfer Learning approach with a Recursive Feature Elimination (MSML-TL-RFE) strategy is proposed to exploit the knowledge extraction power of the model from multiple sources for different days-ahead-prediction, distinguishing itself from the other investigated approaches. In this paper, an abstract concept of a supply chain, with information sharing among retailers, is investigated to show that such concepts can escalate the knowledge transfer ability of a system. A hierarchical two-echelon supply chain model with different attributes is developed to validate the proposed MSML-TL-RFE approach against a few other TL-based forecasting approaches. The feature-rich datasets are then transformed in such a way that they depict a hierarchical supply chain structure, allowing for the effective application of TL for forecasting consumer demand for recently introduced products. Continuing with that idea of information sharing, finding comparable sources for a quick and effective knowledge transfer procedure is investigated, considering all the peculiarities of a certain data set. MSML-TL-REF predictions and other TL-based approaches are analysed by calculating overall supply chain costs. Based on overall supply chain costs under static and dynamic lead time settings, the effectiveness and applicability of the proposed MSML-TL-RFE against traditional forecasting approaches are demonstrated. Incorporating MSML-TL-RFE with three sources improves accuracy, defined as the reciprocal of Root Mean Square Error (RMSE), from 4.83 (no TL) to 5.67 and further increases to 5.76 with additional sources, enabling more accurate predictions and reduced supply chain costs for businesses.
对于供应链从业者来说,预测新产品的未来需求一直是一个挑战,通常是由于缺乏数据。然而,市场调查和通过检查类似的市场产品来发现新产品的行为来提取知识可能是不准确的,并导致错误的结果,最终导致对企业总体成本的错误估计。同时,随着迁移学习(TL)等人工智能方法的进步,通过从其他类似产品的历史数据中寻找知识,更准确地预测新产品的需求,可以减少这种对成本的错误估计。因此,本文研究了几种经典的基于人工智能的TL方法来预测顾客对新产品和新商店的需求。在此基础上,提出了一种基于递归特征消除(MSML-TL-RFE)策略的多源多层迁移学习方法,利用模型从多源中提取不同日前预测的知识能力,区别于其他研究方法。本文研究了具有零售商之间信息共享的供应链的抽象概念,证明了这种概念可以提升系统的知识转移能力。建立了一个具有不同属性的分层两级供应链模型,以验证所提出的MSML-TL-RFE方法与其他几种基于tl的预测方法。然后以这样一种方式转换特征丰富的数据集,即它们描述分层供应链结构,从而允许有效地应用TL来预测消费者对最近推出的产品的需求。考虑到特定数据集的所有特性,继续使用信息共享的思想,研究如何为快速有效的知识转移过程找到可比较的来源。通过计算整体供应链成本来分析MSML-TL-REF预测和其他基于tl的方法。基于静态和动态前置时间设定下的整体供应链成本,验证了MSML-TL-RFE相对于传统预测方法的有效性和适用性。将MSML-TL-RFE与三个来源结合可以提高准确性,定义为均方根误差(RMSE)的倒数,从4.83(无TL)增加到5.67,并进一步增加到5.76,从而实现更准确的预测并降低企业的供应链成本。
{"title":"A multi-source multi-layer-based transfer learning approach for forecasting customer demands of newly launched products","authors":"Supriyo Ahmed ,&nbsp;Ripon K. Chakrabortty ,&nbsp;Daryl L. Essam","doi":"10.1016/j.iswa.2025.200548","DOIUrl":"10.1016/j.iswa.2025.200548","url":null,"abstract":"<div><div>Forecasting the future demand for newly launched products has been challenging for supply chain practitioners, often due to the lack of data. However, market surveys and extracting knowledge by examining similar market products to find the behaviour of a new product can be inaccurate and lead to erroneous results, which ultimately lead to a misestimation of the overall cost of a business. Meanwhile, with the advancement of artificial intelligence (AI) approaches, such as Transfer Learning (TL), this misestimation of cost can be reduced by more accurately forecasting the demand for newly launched products by seeking knowledge from the historical data of other similar products. Consequently, this paper investigates several classical AI-based TL approaches to predict customer demand for new products and stores. Thereafter, a novel <strong>M</strong>ulti-<strong>S</strong>ource <strong>M</strong>ulti-<strong>L</strong>ayer <strong>T</strong>ransfer <strong>L</strong>earning approach with a <strong>R</strong>ecursive <strong>F</strong>eature <strong>E</strong>limination (MSML-TL-RFE) strategy is proposed to exploit the knowledge extraction power of the model from multiple sources for different days-ahead-prediction, distinguishing itself from the other investigated approaches. In this paper, an abstract concept of a supply chain, with information sharing among retailers, is investigated to show that such concepts can escalate the knowledge transfer ability of a system. A hierarchical two-echelon supply chain model with different attributes is developed to validate the proposed MSML-TL-RFE approach against a few other TL-based forecasting approaches. The feature-rich datasets are then transformed in such a way that they depict a hierarchical supply chain structure, allowing for the effective application of TL for forecasting consumer demand for recently introduced products. Continuing with that idea of information sharing, finding comparable sources for a quick and effective knowledge transfer procedure is investigated, considering all the peculiarities of a certain data set. MSML-TL-REF predictions and other TL-based approaches are analysed by calculating overall supply chain costs. Based on overall supply chain costs under static and dynamic lead time settings, the effectiveness and applicability of the proposed MSML-TL-RFE against traditional forecasting approaches are demonstrated. Incorporating MSML-TL-RFE with three sources improves accuracy, defined as the reciprocal of Root Mean Square Error (RMSE), from 4.83 (no TL) to 5.67 and further increases to 5.76 with additional sources, enabling more accurate predictions and reduced supply chain costs for businesses.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"27 ","pages":"Article 200548"},"PeriodicalIF":0.0,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144516953","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A novel binary Stellar Oscillation Optimizer for feature selection optimization problems 一种用于特征选择优化问题的新型双星振荡优化器
Pub Date : 2025-09-01 Epub Date: 2025-07-16 DOI: 10.1016/j.iswa.2025.200558
Ali Rodan , Sharif Naser Makhadmeh , Yousef Sanjalawe , Rizik M.H. Al-Sayyed , Mohammed Azmi Al-Betar
Stellar Oscillation Optimizer (SOO) takes its core inspiration from the study of stellar pulsations, a domain often referred to as asteroseismology which is formulated as an optimization algorithm for continuous domain. In this paper, the Binary version of Stellar Oscillation Optimizer (BSOO) is proposed for Feature Selection (FS) problems. BSOO introduces binary adaptations, including threshold-based encoding, controlled oscillatory movements, and a top-solution influence mechanism. In order to evaluate the BSOO, sixteen FS datasets are used with different numbers of features, samples, and class labels. Seven performance measures are also used, which are: fitness value, number of selected features, accuracy, sensitivity, specificity, Precision, and F-measure. An intensive comparative evaluation against 18 state-of-the-art optimization algorithms using the same datasets has been conducted. The results show that the proposed BSOO version is able to compete well with the other FS-based methods where it is able to overcome several methods and produce the best overall results for some datasets on different measurements. Furthermore, the convergence behavior to show the optimization behavior of BSOO during the search is investigated and visualized. Interestingly, the BSOO is able to provide a suitable trade-off between the global wide-range exploration and local nearby exploitation during the optimization process. This is proved using the statistical Wilcoxon Rank-Sum Test Results. In conclusion, this paper provides a new alternative solution for FS research community that is able to work well for many FS instances and find the optimal solution. The source code of BSOO is publicly available for both MATLAB at: https://www.mathworks.com/matlabcentral/fileexchange/180096-bsoo-binary-stellar-oscillation-optimizer and PYTHON at: https://github.com/AliRodan/BSOO-Binary-Stellar-Oscillation-Optimizer.
恒星振荡优化器(SOO)的核心灵感来自恒星脉动的研究,这一领域通常被称为星震学,它被表述为连续域的优化算法。针对特征选择问题,提出了星振优化器(BSOO)的二进制版本。BSOO引入了二进制自适应,包括基于阈值的编码、可控振荡运动和顶解影响机制。为了评估BSOO,使用了16个具有不同数量的特征,样本和类别标签的FS数据集。还使用了七个性能度量,它们是:适应度值、选择特征的数量、准确性、灵敏度、特异性、精度和f度量。对使用相同数据集的18个最先进的优化算法进行了密集的比较评估。结果表明,所提出的BSOO版本能够很好地与其他基于fs的方法竞争,它能够克服几种方法,并在不同测量的某些数据集上产生最佳的整体结果。此外,研究了BSOO在搜索过程中的收敛行为,并将其可视化。有趣的是,在优化过程中,BSOO能够在全局大范围勘探和局部近距离开采之间提供合适的权衡。使用统计的Wilcoxon秩和检验结果证明了这一点。综上所述,本文为FS研究界提供了一种新的替代解决方案,能够很好地适用于许多FS实例并找到最优解。BSOO的源代码在MATLAB: https://www.mathworks.com/matlabcentral/fileexchange/180096-bsoo-binary-stellar-oscillation-optimizer和PYTHON: https://github.com/AliRodan/BSOO-Binary-Stellar-Oscillation-Optimizer都是公开的。
{"title":"A novel binary Stellar Oscillation Optimizer for feature selection optimization problems","authors":"Ali Rodan ,&nbsp;Sharif Naser Makhadmeh ,&nbsp;Yousef Sanjalawe ,&nbsp;Rizik M.H. Al-Sayyed ,&nbsp;Mohammed Azmi Al-Betar","doi":"10.1016/j.iswa.2025.200558","DOIUrl":"10.1016/j.iswa.2025.200558","url":null,"abstract":"<div><div>Stellar Oscillation Optimizer (SOO) takes its core inspiration from the study of stellar pulsations, a domain often referred to as asteroseismology which is formulated as an optimization algorithm for continuous domain. In this paper, the Binary version of Stellar Oscillation Optimizer (BSOO) is proposed for Feature Selection (FS) problems. BSOO introduces binary adaptations, including threshold-based encoding, controlled oscillatory movements, and a top-solution influence mechanism. In order to evaluate the BSOO, sixteen FS datasets are used with different numbers of features, samples, and class labels. Seven performance measures are also used, which are: fitness value, number of selected features, accuracy, sensitivity, specificity, Precision, and F-measure. An intensive comparative evaluation against 18 state-of-the-art optimization algorithms using the same datasets has been conducted. The results show that the proposed BSOO version is able to compete well with the other FS-based methods where it is able to overcome several methods and produce the best overall results for some datasets on different measurements. Furthermore, the convergence behavior to show the optimization behavior of BSOO during the search is investigated and visualized. Interestingly, the BSOO is able to provide a suitable trade-off between the global wide-range exploration and local nearby exploitation during the optimization process. This is proved using the statistical Wilcoxon Rank-Sum Test Results. In conclusion, this paper provides a new alternative solution for FS research community that is able to work well for many FS instances and find the optimal solution. The source code of BSOO is publicly available for both MATLAB at: <span><span>https://www.mathworks.com/matlabcentral/fileexchange/180096-bsoo-binary-stellar-oscillation-optimizer</span><svg><path></path></svg></span> and PYTHON at: <span><span>https://github.com/AliRodan/BSOO-Binary-Stellar-Oscillation-Optimizer</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"27 ","pages":"Article 200558"},"PeriodicalIF":0.0,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144694572","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Temporal event localization in sports videos via self-supervised proposal generation and cross-modal fusion 基于自监督提案生成和跨模态融合的运动视频时间事件定位
Pub Date : 2025-09-01 Epub Date: 2025-06-14 DOI: 10.1016/j.iswa.2025.200539
Guang Xv , Xingchen Wu
Long-form sports videos present unique challenges for temporally localizing relevant segments described by text queries. In this paper, a novel two-stage method is proposed for text snippet localization in long sports videos, combining efficient retrieval with fine-grained refinement. First, an improved video encoding pipeline with a caching mechanism and a video-centric sampling strategy has been designed to efficiently process long videos. Then, a self-supervised proposal generation module is designed that leverages temporal consistency to generate candidate segments (proposals) with pseudo labels, reducing reliance on exhaustive manual annotation. Our model is trained in two stages: a segment-level discrimination stage that learns to identify short video snippets relevant to a query, followed by an instance-level completeness stage that ensures the entire event described by the query is accurately captured. To effectively fuse visual and textual information, a cross-modal fusion strategy is adopted that combines late fusion for scalable coarse retrieval with targeted cross-modal attention for precise alignment. Experiments on sports video datasets demonstrate that our method outperforms state-of-the-art baselines in both accuracy and efficiency.
长篇体育视频对文本查询描述的相关片段的临时本地化提出了独特的挑战。本文提出了一种将高效检索与细粒度细化相结合的两阶段运动视频文本片段定位方法。首先,设计了一种改进的视频编码管道,该管道具有缓存机制和以视频为中心的采样策略,可以有效地处理长视频。然后,设计了一个自监督提案生成模块,该模块利用时间一致性生成具有伪标签的候选片段(提案),减少了对详尽的手动注释的依赖。我们的模型分为两个阶段进行训练:段级识别阶段,学习识别与查询相关的短视频片段,然后是实例级完整性阶段,确保准确捕获查询所描述的整个事件。为了有效地融合视觉和文本信息,采用了一种跨模态融合策略,将后期融合用于可扩展的粗检索与有针对性的跨模态关注相结合,进行精确对齐。在体育视频数据集上的实验表明,我们的方法在准确性和效率方面都优于最先进的基线。
{"title":"Temporal event localization in sports videos via self-supervised proposal generation and cross-modal fusion","authors":"Guang Xv ,&nbsp;Xingchen Wu","doi":"10.1016/j.iswa.2025.200539","DOIUrl":"10.1016/j.iswa.2025.200539","url":null,"abstract":"<div><div>Long-form sports videos present unique challenges for temporally localizing relevant segments described by text queries. In this paper, a novel two-stage method is proposed for text snippet localization in long sports videos, combining efficient retrieval with fine-grained refinement. First, an improved video encoding pipeline with a caching mechanism and a video-centric sampling strategy has been designed to efficiently process long videos. Then, a self-supervised proposal generation module is designed that leverages temporal consistency to generate candidate segments (proposals) with pseudo labels, reducing reliance on exhaustive manual annotation. Our model is trained in two stages: a segment-level discrimination stage that learns to identify short video snippets relevant to a query, followed by an instance-level completeness stage that ensures the entire event described by the query is accurately captured. To effectively fuse visual and textual information, a cross-modal fusion strategy is adopted that combines late fusion for scalable coarse retrieval with targeted cross-modal attention for precise alignment. Experiments on sports video datasets demonstrate that our method outperforms state-of-the-art baselines in both accuracy and efficiency.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"27 ","pages":"Article 200539"},"PeriodicalIF":0.0,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144514039","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Gradient-enhanced evolutionary multi-objective optimization (GEEMOO): Balancing relevance, learning outcomes, and diversity in educational recommendation systems 梯度增强进化多目标优化(GEEMOO):在教育推荐系统中平衡相关性、学习结果和多样性
IF 4.3 Pub Date : 2025-09-01 Epub Date: 2025-08-10 DOI: 10.1016/j.iswa.2025.200568
Youssef Jdidou , Souhaib Aammou , Hicham Er-radi , Ilias Aarab
The increasing complexity of educational recommendation systems, driven by the need to balance content relevance, learning outcomes, and diversity, demands advanced optimization solutions that overcome the limitations of traditional methods. As educational technology is exponentially improving, multi-objective optimization plays a vital role in adapting learning experiences to individual requirements. This study tackles the Gradient-Enhanced Evolutionary Multi-objective Optimization (GEEMOO) algorithm, which is considered as a hybrid framework that deals with three conflicting objectives: Relevance, Learning Outcomes, and Diversity. GEEMOO associates gradient-based methods for rapid integration with the correlative power of evolutionary strategies to deliver high-quality Pareto-optimal solutions. Extensive experimentation, using real-world datasets, has shown that GEEMOO consistently exceeded benchmark algorithms performance (NSGA-II and MOPSO) across key metrics, achieving greater Hypervolume, Generational Distance, and diversity indicators. While maintaining robust solution diversity, GEEMOO stands as an ideal solution for large-scale educational recommendation systems efficiency, requiring fewer fitness evaluations. GEEMOO showed better performance than NSGA-II and MOPSO in both convergence (Hypervolume: 0.85, Generational Distance: 0.02) and diversity (Spread Indicator: 0.88, Crowding Distance: 0.92). Although it required a bit more runtime (150 seconds compared to 120 seconds for NSGA-II), GEEMOO achieved this with fewer fitness evaluations (50,000 versus 60,000 for NSGA-II), highlighting its computational efficiency. The algorithm successfully balanced conflicting objectives, providing Pareto-optimal solutions that cater to various educational goals. This work traits GEEMOO’s adaptability and credibility to demonstrate how personalized learning models are adjusted, offering a solid groundwork for improving educational technology in both research and practice.
由于需要平衡内容相关性、学习结果和多样性,教育推荐系统的复杂性日益增加,因此需要先进的优化解决方案来克服传统方法的局限性。随着教育技术的指数级发展,多目标优化在使学习体验适应个人需求方面起着至关重要的作用。本研究解决了梯度增强进化多目标优化(GEEMOO)算法,该算法被认为是一个混合框架,处理三个相互冲突的目标:相关性、学习成果和多样性。GEEMOO将基于梯度的方法与进化策略的相关能力相结合,以提供高质量的帕累托最优解决方案。使用真实数据集进行的大量实验表明,GEEMOO在关键指标上始终优于基准算法性能(NSGA-II和MOPSO),实现了更高的Hypervolume、代距和多样性指标。在保持强大的解决方案多样性的同时,GEEMOO是大规模教育推荐系统效率的理想解决方案,需要更少的健身评估。GEEMOO在收敛性(Hypervolume: 0.85,代际距离:0.02)和多样性(Spread Indicator: 0.88,拥挤距离:0.92)方面均优于NSGA-II和MOPSO。虽然它需要更多的运行时间(150秒,而NSGA-II为120秒),但GEEMOO通过更少的适应度评估(50,000次,而NSGA-II为60,000次)实现了这一点,突出了它的计算效率。该算法成功地平衡了相互冲突的目标,提供了满足各种教育目标的帕累托最优解。这项工作突出了GEEMOO的适应性和可信度,展示了个性化学习模式是如何调整的,为在研究和实践中改进教育技术提供了坚实的基础。
{"title":"Gradient-enhanced evolutionary multi-objective optimization (GEEMOO): Balancing relevance, learning outcomes, and diversity in educational recommendation systems","authors":"Youssef Jdidou ,&nbsp;Souhaib Aammou ,&nbsp;Hicham Er-radi ,&nbsp;Ilias Aarab","doi":"10.1016/j.iswa.2025.200568","DOIUrl":"10.1016/j.iswa.2025.200568","url":null,"abstract":"<div><div>The increasing complexity of educational recommendation systems, driven by the need to balance content relevance, learning outcomes, and diversity, demands advanced optimization solutions that overcome the limitations of traditional methods. As educational technology is exponentially improving, multi-objective optimization plays a vital role in adapting learning experiences to individual requirements. This study tackles the Gradient-Enhanced Evolutionary Multi-objective Optimization (GEEMOO) algorithm, which is considered as a hybrid framework that deals with three conflicting objectives: Relevance, Learning Outcomes, and Diversity. GEEMOO associates gradient-based methods for rapid integration with the correlative power of evolutionary strategies to deliver high-quality Pareto-optimal solutions. Extensive experimentation, using real-world datasets, has shown that GEEMOO consistently exceeded benchmark algorithms performance (NSGA-II and MOPSO) across key metrics, achieving greater Hypervolume, Generational Distance, and diversity indicators. While maintaining robust solution diversity, GEEMOO stands as an ideal solution for large-scale educational recommendation systems efficiency, requiring fewer fitness evaluations. GEEMOO showed better performance than NSGA-II and MOPSO in both convergence (Hypervolume: 0.85, Generational Distance: 0.02) and diversity (Spread Indicator: 0.88, Crowding Distance: 0.92). Although it required a bit more runtime (150 seconds compared to 120 seconds for NSGA-II), GEEMOO achieved this with fewer fitness evaluations (50,000 versus 60,000 for NSGA-II), highlighting its computational efficiency. The algorithm successfully balanced conflicting objectives, providing Pareto-optimal solutions that cater to various educational goals. This work traits GEEMOO’s adaptability and credibility to demonstrate how personalized learning models are adjusted, offering a solid groundwork for improving educational technology in both research and practice.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"27 ","pages":"Article 200568"},"PeriodicalIF":4.3,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144828696","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LWR-Net: Learning without retraining for scalable multi-task adaptation and domain-agnostic generalisation LWR-Net:无需再训练的可扩展多任务适应和领域不可知泛化学习
IF 4.3 Pub Date : 2025-09-01 Epub Date: 2025-08-11 DOI: 10.1016/j.iswa.2025.200567
Haider A. Alwzwazy , Laith Alzubaidi , Zehui Zhao , Ahmed Saihood , Sabah Abdulazeez Jebur , Mohamed Manoufali , Omar Alnaseri , Jose Santamaria , Yuantong Gu
In recent years, deep learning-based multi-class and multi-task classification have gained significant attention across various domains of computer vision. However, current approaches often struggle to incorporate new classes efficiently due to the computational burden of retraining large neural networks from scratch. This limitation poses a significant obstacle to the deployment of deep learning models in real-world intelligent systems. Although continual learning has been proposed to overcome this challenge, it remains constrained by catastrophic forgetting. To address these limitations, this study introduces a new framework called Learning Without Retraining (LWR-Net), developed for multi-class and multi-task adaptation, allowing networks to adapt to new classes with minimal training requirements. Specifically, LWR-Net incorporates four key components: (i) task-guided self-supervised learning with a dual-attention mechanism to enhance feature generalisation and selection; (ii) task-based model fusion to improve feature representation and generalisation; (iii) multi-task learning to generalise classifiers across diverse tasks; and (iv) decision fusion of multiple classifiers to improve overall performance and reduce the likelihood of misclassification. LWR-Net was evaluated across diverse tasks to demonstrate its effectiveness in integrating new data, classes, or tasks. These include: (i) a medical case study detecting abnormalities in five distinct bone structures; (ii) a surveillance case study detecting violence in three different settings; and (iii) a geology case study identifying lateral changes in soil compaction using ground-penetrating radar across two datasets. The results show that LWR-Net achieves state-of-the-art performance across all three scenarios, successfully accommodates new learning objectives while preserving performance, eliminating the need for complete retraining cycles. Moreover, the use of gradient-weighted class activation mapping (Grad-CAM) confirmed that the models focused on relevant regions of interest. LWR-Net offers several benefits, including improved generalisation, enhanced performance, and the capacity to train on new data without catastrophic failures. The source code is publicly available at: https://github.com/LaithAlzubaidi/Learning-to-Adapt.
近年来,基于深度学习的多类和多任务分类在计算机视觉的各个领域得到了广泛的关注。然而,由于从头开始重新训练大型神经网络的计算负担,目前的方法常常难以有效地合并新类。这一限制对在现实世界的智能系统中部署深度学习模型构成了重大障碍。尽管持续学习已经被提出来克服这一挑战,但它仍然受到灾难性遗忘的限制。为了解决这些限制,本研究引入了一个名为“无需再培训的学习”(LWR-Net)的新框架,该框架是为多类和多任务适应而开发的,允许网络以最小的培训要求适应新类。具体而言,LWR-Net包含四个关键组成部分:(i)任务引导的自监督学习与双注意机制,以增强特征的泛化和选择;(ii)基于任务的模型融合,改进特征表示和泛化;(iii)多任务学习,在不同的任务中泛化分类器;(iv)多分类器的决策融合,提高整体性能,降低误分类的可能性。对LWR-Net进行了跨不同任务的评估,以证明其在集成新数据、类或任务方面的有效性。其中包括:(i)一项医学案例研究,发现五种不同骨骼结构的异常情况;(ii)在三种不同情况下发现暴力的监测案例研究;(iii)地质案例研究,利用探地雷达在两个数据集上识别土壤压实的横向变化。结果表明,LWR-Net在所有三种情况下都达到了最先进的性能,成功地适应了新的学习目标,同时保持了性能,消除了对完整再训练周期的需要。此外,使用梯度加权类激活映射(Grad-CAM)证实了模型专注于感兴趣的相关区域。LWR-Net提供了几个好处,包括改进的泛化、增强的性能以及在没有灾难性故障的情况下对新数据进行训练的能力。源代码可以在:https://github.com/LaithAlzubaidi/Learning-to-Adapt上公开获得。
{"title":"LWR-Net: Learning without retraining for scalable multi-task adaptation and domain-agnostic generalisation","authors":"Haider A. Alwzwazy ,&nbsp;Laith Alzubaidi ,&nbsp;Zehui Zhao ,&nbsp;Ahmed Saihood ,&nbsp;Sabah Abdulazeez Jebur ,&nbsp;Mohamed Manoufali ,&nbsp;Omar Alnaseri ,&nbsp;Jose Santamaria ,&nbsp;Yuantong Gu","doi":"10.1016/j.iswa.2025.200567","DOIUrl":"10.1016/j.iswa.2025.200567","url":null,"abstract":"<div><div>In recent years, deep learning-based multi-class and multi-task classification have gained significant attention across various domains of computer vision. However, current approaches often struggle to incorporate new classes efficiently due to the computational burden of retraining large neural networks from scratch. This limitation poses a significant obstacle to the deployment of deep learning models in real-world intelligent systems. Although continual learning has been proposed to overcome this challenge, it remains constrained by catastrophic forgetting. To address these limitations, this study introduces a new framework called Learning Without Retraining (LWR-Net), developed for multi-class and multi-task adaptation, allowing networks to adapt to new classes with minimal training requirements. Specifically, LWR-Net incorporates four key components: (i) task-guided self-supervised learning with a dual-attention mechanism to enhance feature generalisation and selection; (ii) task-based model fusion to improve feature representation and generalisation; (iii) multi-task learning to generalise classifiers across diverse tasks; and (iv) decision fusion of multiple classifiers to improve overall performance and reduce the likelihood of misclassification. LWR-Net was evaluated across diverse tasks to demonstrate its effectiveness in integrating new data, classes, or tasks. These include: (i) a medical case study detecting abnormalities in five distinct bone structures; (ii) a surveillance case study detecting violence in three different settings; and (iii) a geology case study identifying lateral changes in soil compaction using ground-penetrating radar across two datasets. The results show that LWR-Net achieves state-of-the-art performance across all three scenarios, successfully accommodates new learning objectives while preserving performance, eliminating the need for complete retraining cycles. Moreover, the use of gradient-weighted class activation mapping (Grad-CAM) confirmed that the models focused on relevant regions of interest. LWR-Net offers several benefits, including improved generalisation, enhanced performance, and the capacity to train on new data without catastrophic failures. The source code is publicly available at: <span><span>https://github.com/LaithAlzubaidi/Learning-to-Adapt</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"27 ","pages":"Article 200567"},"PeriodicalIF":4.3,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144828697","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Intelligent Systems with Applications
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1