首页 > 最新文献

Computational Intelligence最新文献

英文 中文
IoMT-Based Multiple Disease Detection From Medical Images Using Vision Transformer-Based Multi-Scale Residual DenseNet With Hybrid Heuristic Strategies 基于视觉变换的多尺度残差密度网混合启发式医学图像多重疾病检测
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-14 DOI: 10.1111/coin.70136
Karunakar Chappidi, L. N. B. Srinivas

In the Internet of Medical Things (IoMT), medical images play a vital role in healthcare systems by enabling accurate diagnosis, effective treatment planning, and continuous monitoring of various diseases. Deep learning models have strong potential to facilitate early detection by recognizing patterns in medical images. Although deep learning systems are more advanced, ambiguity continues to act as a major obstacle to progress. Ambiguity typically arises in areas such as accuracy, interpretability, and generalization. The challenge lies in balancing high performance with transparent and understandable decision-making processes. This uncertainty can hinder trust computation and practical deployment, especially in critical applications. Moreover, these models face challenges with data handling and often lack robustness, which affects their reliability. To address the above-mentioned issues, a novel multi-disease detection framework is proposed using a deep learning approach. The input medical images are gathered from online datasets using the IoMT-enabled devices. To detect different kinds of diseases, an adaptive vision transformer-based multi-scale residual dense network (AViT-MRDNet) is developed. This model processes the input images to classify four types of diseases including skin lesions, brain tumors, lung tumors, and breast cancer. Here, the hybrid bonobo with rat swarm optimizer (HB-RSO) is developed to optimize the parameters. Finally, the proposed framework produces the disease detection results. The overall performance of the implemented framework is verified by using various validation metrics such as precision, accuracy, false discovery rate, sensitivity, F1-score, false positive rate, specificity, Mathew correlation coefficient, false negative rate, and negative predictive value to showcase the effectiveness in detecting the multi-disease. It attains an accuracy of 96.022, sensitivity of 95.164, specificity of 96.880, and precision of 96.825. Thus, these findings prove that the introduced framework has enhanced the system performance. The hybrid optimization algorithm in the proposed approach offers a robust solution for doctors and healthcare professionals to predict multiple diseases in patients. Therefore, it shows that the developed model is superior and delivers more reliable performance to other frameworks. The result confirmed that the proposed model can accurately and precisely detect the multi-disease and help the patients for early diagnosis. It leverages the strength of the vision transformer and the residual densenet to enhance the system performance.

在医疗物联网(IoMT)中,医学图像在医疗保健系统中发挥着至关重要的作用,可以实现准确的诊断、有效的治疗计划和对各种疾病的持续监测。深度学习模型具有很强的潜力,可以通过识别医学图像中的模式来促进早期检测。尽管深度学习系统更加先进,但模糊性仍然是进步的主要障碍。歧义通常出现在准确性、可解释性和泛化等方面。挑战在于如何在高绩效与透明和可理解的决策过程之间取得平衡。这种不确定性会阻碍信任计算和实际部署,特别是在关键应用程序中。此外,这些模型面临着数据处理方面的挑战,往往缺乏鲁棒性,从而影响了它们的可靠性。为了解决上述问题,本文提出了一种基于深度学习的多疾病检测框架。输入的医学图像是使用支持iomt的设备从在线数据集收集的。为了检测不同类型的疾病,提出了一种基于自适应视觉变换的多尺度残差密集网络(AViT-MRDNet)。该模型对输入的图像进行处理,对皮肤病变、脑肿瘤、肺癌和乳腺癌等四种疾病进行分类。本文采用混合倭黑猩猩与大鼠群优化器(HB-RSO)进行参数优化。最后,该框架生成疾病检测结果。采用精密度、准确度、假发现率、灵敏度、f1评分、假阳性率、特异性、马修相关系数、假阴性率、阴性预测值等多种验证指标对所实现框架的总体性能进行验证,以展示其在检测多种疾病方面的有效性。准确度为96.022,灵敏度为95.164,特异度为96.880,精密度为96.825。因此,这些发现证明了引入的框架提高了系统的性能。该方法中的混合优化算法为医生和医疗保健专业人员预测患者的多种疾病提供了一个强大的解决方案。由此可见,所开发的模型比其他框架具有更优越的性能和更可靠的性能。结果表明,该模型能够准确准确地检测出多种疾病,有助于患者早期诊断。它利用视觉变压器的强度和残余密度来提高系统的性能。
{"title":"IoMT-Based Multiple Disease Detection From Medical Images Using Vision Transformer-Based Multi-Scale Residual DenseNet With Hybrid Heuristic Strategies","authors":"Karunakar Chappidi,&nbsp;L. N. B. Srinivas","doi":"10.1111/coin.70136","DOIUrl":"https://doi.org/10.1111/coin.70136","url":null,"abstract":"<div>\u0000 \u0000 <p>In the Internet of Medical Things (IoMT), medical images play a vital role in healthcare systems by enabling accurate diagnosis, effective treatment planning, and continuous monitoring of various diseases. Deep learning models have strong potential to facilitate early detection by recognizing patterns in medical images. Although deep learning systems are more advanced, ambiguity continues to act as a major obstacle to progress. Ambiguity typically arises in areas such as accuracy, interpretability, and generalization. The challenge lies in balancing high performance with transparent and understandable decision-making processes. This uncertainty can hinder trust computation and practical deployment, especially in critical applications. Moreover, these models face challenges with data handling and often lack robustness, which affects their reliability. To address the above-mentioned issues, a novel multi-disease detection framework is proposed using a deep learning approach. The input medical images are gathered from online datasets using the IoMT-enabled devices. To detect different kinds of diseases, an adaptive vision transformer-based multi-scale residual dense network (AViT-MRDNet) is developed. This model processes the input images to classify four types of diseases including skin lesions, brain tumors, lung tumors, and breast cancer. Here, the hybrid bonobo with rat swarm optimizer (HB-RSO) is developed to optimize the parameters. Finally, the proposed framework produces the disease detection results. The overall performance of the implemented framework is verified by using various validation metrics such as precision, accuracy, false discovery rate, sensitivity, <i>F</i>1-score, false positive rate, specificity, Mathew correlation coefficient, false negative rate, and negative predictive value to showcase the effectiveness in detecting the multi-disease. It attains an accuracy of 96.022, sensitivity of 95.164, specificity of 96.880, and precision of 96.825. Thus, these findings prove that the introduced framework has enhanced the system performance. The hybrid optimization algorithm in the proposed approach offers a robust solution for doctors and healthcare professionals to predict multiple diseases in patients. Therefore, it shows that the developed model is superior and delivers more reliable performance to other frameworks. The result confirmed that the proposed model can accurately and precisely detect the multi-disease and help the patients for early diagnosis. It leverages the strength of the vision transformer and the residual densenet to enhance the system performance.</p>\u0000 </div>","PeriodicalId":55228,"journal":{"name":"Computational Intelligence","volume":"41 5","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-10-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145317004","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fuzzy-HAGRN: Fuzzy-Based Hierarchical Attention Gated Recurrent Network for Sentiment Classification and Review Rate Prediction Fuzzy-HAGRN:基于模糊的层次注意门控递归网络,用于情感分类和评论率预测
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-13 DOI: 10.1111/coin.70147
Rashmi Thakur, Anil Vasoya, Manoj Chavan, Payel Saha, Sanjeev Ghosh, Sonia Behra

Online reviews are essential for users to make purchasing decisions over any products or films. Reviews found on e-commerce websites often closely align with ratings and enhance the learning experience for potential buyers. The customer review information is of huge interest to companies as well as customers. The information from customer feedback is not only valuable for consumers but also provides valuable feedback to companies. However, since customer reviews are usually presented as unstructured free text, extracting meaningful ratings from user opinions can be quite challenging. To overcome this issue, an effectual system is modeled for sentiment classification and review rating prediction utilizing reviews named fuzzy-based hierarchical attention gated recurrent network (Fuzzy-HAGRN). At first, the input reviews are fed to Bidirectional Transformers for Language Understanding (BERT) tokenization, and after that, feature extraction is executed for the extraction of several essential features. Then, Fuzzy-HAGRN is utilized for sentiment classification and review rate prediction. Here, the Fuzzy-HAGRN method is developed by integrating the fuzzy concept, hierarchical attention network (HAN) as well as gated recurrent network (GRU). Finally, rating prediction is performed using the same Fuzzy-HAGRN. Fuzzy-HAGRN has obtained a precision of 92.20%, recall of 91.80%, and F1-score of 91.50%.

在线评论对于用户决定购买任何产品或电影都是至关重要的。电子商务网站上的评论通常与评级密切相关,可以增强潜在买家的学习体验。客户评论信息对于公司和客户来说都是非常重要的。顾客反馈的信息不仅对消费者有价值,对企业也有价值。然而,由于用户评论通常以非结构化的自由文本形式呈现,因此从用户意见中提取有意义的评级可能相当具有挑战性。为了克服这一问题,利用基于模糊的分层注意力门控递归网络(Fuzzy-HAGRN)建立了一个有效的情感分类和评论评级预测系统模型。首先,将输入评论馈送到用于语言理解的双向转换(BERT)标记化,然后执行特征提取以提取几个基本特征。然后,利用Fuzzy-HAGRN进行情感分类和评论率预测。本文将模糊概念、层次注意网络(HAN)和门控循环网络(GRU)相结合,提出了fuzzy - hagrn方法。最后,使用相同的Fuzzy-HAGRN进行评级预测。Fuzzy-HAGRN的准确率为92.20%,召回率为91.80%,f1得分为91.50%。
{"title":"Fuzzy-HAGRN: Fuzzy-Based Hierarchical Attention Gated Recurrent Network for Sentiment Classification and Review Rate Prediction","authors":"Rashmi Thakur,&nbsp;Anil Vasoya,&nbsp;Manoj Chavan,&nbsp;Payel Saha,&nbsp;Sanjeev Ghosh,&nbsp;Sonia Behra","doi":"10.1111/coin.70147","DOIUrl":"https://doi.org/10.1111/coin.70147","url":null,"abstract":"<div>\u0000 \u0000 <p>Online reviews are essential for users to make purchasing decisions over any products or films. Reviews found on e-commerce websites often closely align with ratings and enhance the learning experience for potential buyers. The customer review information is of huge interest to companies as well as customers. The information from customer feedback is not only valuable for consumers but also provides valuable feedback to companies. However, since customer reviews are usually presented as unstructured free text, extracting meaningful ratings from user opinions can be quite challenging. To overcome this issue, an effectual system is modeled for sentiment classification and review rating prediction utilizing reviews named fuzzy-based hierarchical attention gated recurrent network (Fuzzy-HAGRN). At first, the input reviews are fed to Bidirectional Transformers for Language Understanding (BERT) tokenization, and after that, feature extraction is executed for the extraction of several essential features. Then, Fuzzy-HAGRN is utilized for sentiment classification and review rate prediction. Here, the Fuzzy-HAGRN method is developed by integrating the fuzzy concept, hierarchical attention network (HAN) as well as gated recurrent network (GRU). Finally, rating prediction is performed using the same Fuzzy-HAGRN. Fuzzy-HAGRN has obtained a precision of 92.20%, recall of 91.80%, and <i>F</i>1-score of 91.50%.</p>\u0000 </div>","PeriodicalId":55228,"journal":{"name":"Computational Intelligence","volume":"41 5","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145316724","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing Cloud Security by Performing Deduplication Using Serial Cascaded Autoencoder With GRU and Optimal Key-Based Data Sanitization 通过使用串行级联自编码器与GRU和最优的基于密钥的数据清理执行重复数据删除增强云安全
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-13 DOI: 10.1111/coin.70140
J. K. Periasamy, Chin-Shiuh Shieh, Mong-Fong Horng

De-duplication is critically important for cloud computing since it permits the detection of repeated data within the cloud system under fewer resources and expenses. De-duplication removes unnecessary data from the cloud centers, which helps to identify the appropriate owner of cloud material. Each piece of data saved in the cloud is owned by a large number of cloud users, even though it contains just a single copy of the data. The dynamic nature of the cloud resources is not handled by the prior deduplication models and the existing models require more computing power for accurately determining the presence of duplicate files in the cloud system. In addition, the prior models split the files into chunks for determining the similar files in the cloud system which affects the quality of the data. To conquer these difficulties, an adaptive deep learning-based data deduplication model is developed using an optimization algorithm. The main innovation of the proposed research is to rapidly detect duplicate records in the cloud data and also provide high-level security while maintaining the operational efficiency of the cloud system. The proposed model acts as an efficient attack resistance system and it also ensures the data availability of the cloud system more rapidly. This data deduplication implies in detecting and examining the patterns inside records of information to precisely notice and eliminate repeated identical information. Hence, the data connected to the input pattern is given to the Serial Cascaded Autoencoder with Gated Recurrent Unit (SCA-GRU) for the deduplication process. After deduplication, the unnecessary data are removed for the precise consumption of resources to store exclusive data. To maintain the security of data, the optimal key-based data sanitization process is performed, in which the key is optimally generated with the aid of a Mutated Fitness-Based Krill Herd Optimization Algorithm (MF-KHO). This encoded data is then safely kept in the cloud, which protects the data from illegal access and possible defense breaches. The outcome of the suggested approach is validated with the previous data deduplication system to show the efficiency of the developed model. The experimental results showed that the recommended deduplication approach reaches an accuracy of 95.37%. Through efficient data deduplication, the storage requirement of the data is greatly reduced, which facilitates cost reduction and resource optimization within the cloud system and also the storage capacity utilization of the cloud system is greatly improved.

重复数据删除对云计算至关重要,因为它允许以更少的资源和费用检测云系统中的重复数据。重复数据删除可以从云中心删除不必要的数据,这有助于确定云数据的适当所有者。保存在云中的每一段数据都由大量的云用户拥有,即使它只包含数据的一个副本。以前的重复数据删除模型无法处理云资源的动态性,现有模型需要更多的计算能力才能准确地确定云系统中是否存在重复文件。此外,先前的模型将文件分割成块,以确定云系统中影响数据质量的相似文件。为了克服这些困难,利用优化算法开发了一种基于自适应深度学习的重复数据删除模型。本研究的主要创新点在于快速检测云数据中的重复记录,并在保持云系统运行效率的同时提供高水平的安全性。该模型作为一种高效的抗攻击系统,保证了云系统数据的快速可用性。这种数据重复意味着检测和检查信息记录中的模式,以精确地注意和消除重复的相同信息。因此,连接到输入模式的数据被提供给带有门控循环单元的串行级联自编码器(SCA-GRU)用于重复数据删除过程。重复数据删除后,可以删除不需要的数据,从而精确地使用资源来存储独占的数据。为了保证数据的安全性,执行了基于密钥的最佳数据处理过程,其中密钥是在基于突变适应度的磷虾群优化算法(MF-KHO)的帮助下最优生成的。这些编码后的数据被安全地保存在云端,从而保护数据免受非法访问和可能的防御漏洞。最后,将该方法的结果与之前的重复数据删除系统进行了验证,以显示所开发模型的有效性。实验结果表明,推荐的重复数据删除方法准确率达到95.37%。通过高效的重复数据删除,大大降低了数据的存储需求,有利于云系统内部的成本降低和资源优化,也大大提高了云系统的存储容量利用率。
{"title":"Enhancing Cloud Security by Performing Deduplication Using Serial Cascaded Autoencoder With GRU and Optimal Key-Based Data Sanitization","authors":"J. K. Periasamy,&nbsp;Chin-Shiuh Shieh,&nbsp;Mong-Fong Horng","doi":"10.1111/coin.70140","DOIUrl":"https://doi.org/10.1111/coin.70140","url":null,"abstract":"<div>\u0000 \u0000 <p>De-duplication is critically important for cloud computing since it permits the detection of repeated data within the cloud system under fewer resources and expenses. De-duplication removes unnecessary data from the cloud centers, which helps to identify the appropriate owner of cloud material. Each piece of data saved in the cloud is owned by a large number of cloud users, even though it contains just a single copy of the data. The dynamic nature of the cloud resources is not handled by the prior deduplication models and the existing models require more computing power for accurately determining the presence of duplicate files in the cloud system. In addition, the prior models split the files into chunks for determining the similar files in the cloud system which affects the quality of the data. To conquer these difficulties, an adaptive deep learning-based data deduplication model is developed using an optimization algorithm. The main innovation of the proposed research is to rapidly detect duplicate records in the cloud data and also provide high-level security while maintaining the operational efficiency of the cloud system. The proposed model acts as an efficient attack resistance system and it also ensures the data availability of the cloud system more rapidly. This data deduplication implies in detecting and examining the patterns inside records of information to precisely notice and eliminate repeated identical information. Hence, the data connected to the input pattern is given to the Serial Cascaded Autoencoder with Gated Recurrent Unit (SCA-GRU) for the deduplication process. After deduplication, the unnecessary data are removed for the precise consumption of resources to store exclusive data. To maintain the security of data, the optimal key-based data sanitization process is performed, in which the key is optimally generated with the aid of a Mutated Fitness-Based Krill Herd Optimization Algorithm (MF-KHO). This encoded data is then safely kept in the cloud, which protects the data from illegal access and possible defense breaches. The outcome of the suggested approach is validated with the previous data deduplication system to show the efficiency of the developed model. The experimental results showed that the recommended deduplication approach reaches an accuracy of 95.37%. Through efficient data deduplication, the storage requirement of the data is greatly reduced, which facilitates cost reduction and resource optimization within the cloud system and also the storage capacity utilization of the cloud system is greatly improved.</p>\u0000 </div>","PeriodicalId":55228,"journal":{"name":"Computational Intelligence","volume":"41 5","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145316746","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mobile Application for Real-Time Fabric Pattern Classification to Assist Visually Impaired and Blind: A Proof-of-Concept Implementation 帮助视障人士和盲人的实时织物模式分类移动应用程序:概念验证实现
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-12 DOI: 10.1111/coin.70141
Sumeet Saurav, Seema Choudhary, Sanjay Singh

Visual impairment has a drastic impact on the psychological and cognitive well-being of individuals. Recent progress in advanced assistive technologies (AATs) has emerged as an essential tool to mitigate the adverse impact of blindness and enhance the quality of life of visually impaired persons (VIPs). Like generic object identification, the VIPs face difficulties in identifying their garments. Such a limitation severely impacts their identity as they cannot select dresses according to their preferences for different contexts and occasions. To this end, in this paper, we present a proof-of-concept (POC) implementation of a mobile application for real-time fabric pattern classification to assist VIPs in selecting the cloth with fabric patterns of their choice. The proposed framework uses a robust and compute-efficient convolutional neural network (CNN) named FabricNet to classify four types of fabric patterns (lattice, printed, solid, and stripe). The designed FabricNet model uses efficient feature enhancement (EFE), efficient feature refinement (EFR), and enhanced feature fusion (EFF) blocks to extract discriminative texture features from the fabric pattern images. We evaluated the performance of the proposed FabricNet on a recent open-source fabric image dataset. Compared to the existing baseline, the FabricNet model, with 1.64 M parameters, 6.25 MB memory storage size, and 4.09 GFLOPs, attained state-of-the-art classification accuracy with reduced model storage size and competitive computation time. Finally, we optimized the trained model and integrated it with a mobile application developed using the Android Studio software development kit (SDK) for real-time classification of fabric images. The designed mobile application on an Android mobile phone uses its back camera to capture garment images and provides predicted fabric type using audio feedback. Thus, it can assist VIPs in selecting clothing with fabric patterns of their choice in real time.

视觉障碍对个体的心理和认知健康有着巨大的影响。先进辅助技术(AATs)的最新进展已成为减轻失明不利影响和提高视障者(vip)生活质量的重要工具。就像一般的物体识别一样,贵宾们在识别自己的衣服时也面临着困难。这种限制严重影响了她们的身份,因为她们不能根据自己的喜好在不同的场合和场合选择服装。为此,在本文中,我们提出了一个用于实时织物图案分类的移动应用程序的概念验证(POC)实现,以帮助贵宾选择具有其选择的织物图案的织物。所提出的框架使用一个名为FabricNet的鲁棒且计算效率高的卷积神经网络(CNN)来对四种类型的织物图案(格子、印花、实体和条纹)进行分类。设计的FabricNet模型采用高效特征增强(EFE)、高效特征细化(EFR)和增强特征融合(EFF)块从织物图案图像中提取判别性纹理特征。我们在最近的开源织物图像数据集上评估了所提出的FabricNet的性能。与现有基线相比,具有1.64 M参数,6.25 MB内存存储大小和4.09 GFLOPs的FabricNet模型在减少模型存储大小和具有竞争力的计算时间的情况下获得了最先进的分类精度。最后,我们对训练好的模型进行了优化,并将其与Android Studio软件开发工具包(SDK)开发的移动应用程序集成,用于织物图像的实时分类。设计的移动应用程序安装在Android手机上,使用后置摄像头捕捉服装图像,并通过音频反馈提供预测的面料类型。因此,它可以帮助vip实时选择自己选择的面料图案的衣服。
{"title":"Mobile Application for Real-Time Fabric Pattern Classification to Assist Visually Impaired and Blind: A Proof-of-Concept Implementation","authors":"Sumeet Saurav,&nbsp;Seema Choudhary,&nbsp;Sanjay Singh","doi":"10.1111/coin.70141","DOIUrl":"https://doi.org/10.1111/coin.70141","url":null,"abstract":"<div>\u0000 \u0000 <p>Visual impairment has a drastic impact on the psychological and cognitive well-being of individuals. Recent progress in advanced assistive technologies (AATs) has emerged as an essential tool to mitigate the adverse impact of blindness and enhance the quality of life of visually impaired persons (VIPs). Like generic object identification, the VIPs face difficulties in identifying their garments. Such a limitation severely impacts their identity as they cannot select dresses according to their preferences for different contexts and occasions. To this end, in this paper, we present a proof-of-concept (POC) implementation of a mobile application for real-time fabric pattern classification to assist VIPs in selecting the cloth with fabric patterns of their choice. The proposed framework uses a robust and compute-efficient convolutional neural network (CNN) named FabricNet to classify four types of fabric patterns (lattice, printed, solid, and stripe). The designed FabricNet model uses efficient feature enhancement (EFE), efficient feature refinement (EFR), and enhanced feature fusion (EFF) blocks to extract discriminative texture features from the fabric pattern images. We evaluated the performance of the proposed FabricNet on a recent open-source fabric image dataset. Compared to the existing baseline, the FabricNet model, with 1.64 M parameters, 6.25 MB memory storage size, and 4.09 GFLOPs, attained state-of-the-art classification accuracy with reduced model storage size and competitive computation time. Finally, we optimized the trained model and integrated it with a mobile application developed using the Android Studio software development kit (SDK) for real-time classification of fabric images. The designed mobile application on an Android mobile phone uses its back camera to capture garment images and provides predicted fabric type using audio feedback. Thus, it can assist VIPs in selecting clothing with fabric patterns of their choice in real time.</p>\u0000 </div>","PeriodicalId":55228,"journal":{"name":"Computational Intelligence","volume":"41 5","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145316816","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Novel Nonlinearity Extracting Method of Diverse Music Signals Based on Chaotic Techniques for Musical Processing System 音乐处理系统中基于混沌技术的多种音乐信号非线性提取新方法
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-11 DOI: 10.1111/coin.70138
Xueqing Huang, Na Long, Xiaolei Yang

Diverse musical styles are crucial ways for human beings to represent their emotions and interact with each other, whereas the essentials of musical signals are a time-lagged nonlinear dynamical system and their nonlinearity is difficult to analyze by conventional approaches. In this paper, the music is firstly framed depending on the subsections of its structure, then the Lyapunov exponent and the correlation dimension of the music signal are computationally analyzed, which reveals that the internal construction of the music signal is sophisticated with weak chaotic features. By retrieving the local characteristics of the music signal and extrapolating its holistic characteristics, the nonlinearity of the signal rendered by diverse musical styles also has a distinguishable difference. It is observed from the experiments that the maximum Lyapunov exponent of music characterized as “happy” and “relaxing” reaches 0.23, while the range of fluctuations in the correlation dimensions spans from 3.2 to 5.7. Furthermore, a discrepancy of 4.1 is noted in the correlation dimensions of music classified as “loud” and “uplifting,” indicative of the intricate nature of music signals' internal structures and the attenuation of chaotic characteristics. The M5 model exhibits an accuracy of 91.26% for classical music, representing a 2.9% enhancement over conventional methodologies. According to the aforementioned chaotic analysis, the originally designed nonlinearity extracting pattern for diverse music signals in the musical recognizing system demonstrates excellent performance.

多样的音乐风格是人类表达情感和相互影响的重要方式,而音乐信号的本质是一个时滞的非线性动力系统,其非线性难以用常规方法分析。本文首先根据音乐的结构分段对其进行框架化,然后对音乐信号的李雅普诺夫指数和相关维数进行计算分析,揭示了音乐信号的内部结构复杂,具有弱混沌特征。通过检索音乐信号的局部特征并外推其整体特征,不同音乐风格所呈现的信号的非线性也具有明显的差异。从实验中可以观察到,以“快乐”和“放松”为特征的音乐的最大Lyapunov指数达到0.23,而相关维度的波动范围在3.2到5.7之间。此外,在分类为“响亮”和“振奋”的音乐的相关维度中,差异为4.1,这表明音乐信号内部结构的复杂性和混沌特征的衰减。M5模型对古典音乐的准确率为91.26%,比传统方法提高了2.9%。根据前面的混沌分析,原始设计的音乐识别系统中各种音乐信号的非线性提取模式表现出了良好的性能。
{"title":"Novel Nonlinearity Extracting Method of Diverse Music Signals Based on Chaotic Techniques for Musical Processing System","authors":"Xueqing Huang,&nbsp;Na Long,&nbsp;Xiaolei Yang","doi":"10.1111/coin.70138","DOIUrl":"https://doi.org/10.1111/coin.70138","url":null,"abstract":"<div>\u0000 \u0000 <p>Diverse musical styles are crucial ways for human beings to represent their emotions and interact with each other, whereas the essentials of musical signals are a time-lagged nonlinear dynamical system and their nonlinearity is difficult to analyze by conventional approaches. In this paper, the music is firstly framed depending on the subsections of its structure, then the Lyapunov exponent and the correlation dimension of the music signal are computationally analyzed, which reveals that the internal construction of the music signal is sophisticated with weak chaotic features. By retrieving the local characteristics of the music signal and extrapolating its holistic characteristics, the nonlinearity of the signal rendered by diverse musical styles also has a distinguishable difference. It is observed from the experiments that the maximum Lyapunov exponent of music characterized as “happy” and “relaxing” reaches 0.23, while the range of fluctuations in the correlation dimensions spans from 3.2 to 5.7. Furthermore, a discrepancy of 4.1 is noted in the correlation dimensions of music classified as “loud” and “uplifting,” indicative of the intricate nature of music signals' internal structures and the attenuation of chaotic characteristics. The M5 model exhibits an accuracy of 91.26% for classical music, representing a 2.9% enhancement over conventional methodologies. According to the aforementioned chaotic analysis, the originally designed nonlinearity extracting pattern for diverse music signals in the musical recognizing system demonstrates excellent performance.</p>\u0000 </div>","PeriodicalId":55228,"journal":{"name":"Computational Intelligence","volume":"41 5","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145272889","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Seasonality-Aware, Positional, and Topological-Guided GNN (SPT-GNN) for Movie Recommendation 季节性感知、位置和拓扑引导的电影推荐GNN (SPT-GNN)
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-11 DOI: 10.1111/coin.70148
Cevher Özden, Alper Özcan

There has been an increasing interest in using GNNs to build recommender systems as they enable the representation of complex relationships between users and items through knowledge graph embeddings. However, most of the knowledge-graph-based systems focus only on ratings or reviews to build relationships. This prevents a comprehensive understanding of structural and positional information within graph data as well as user preferences that can change in time, as well. In order to address these issues, this paper aims to propose an advanced end-to-end Graph Neural Network architecture that significantly enhances recommendation system capabilities through the integration of state-of-the-art embedding techniques, knowledge graph frameworks, and transfer learning strategies. Incorporating positional encoding and topological feature extraction, the proposed model captures intricate user–item relationships and offers a robust representation that surpasses current approaches. A pretrained encoder facilitates knowledge transfer, effectively bridging domain gaps and amplifying prediction accuracy. Comprehensive evaluations against established baseline models reveal that our architecture has demonstrated enhanced accuracy, precision, and overall robustness. These results highlight the efficacy of combining knowledge graphs, sophisticated embedding strategies, and cross-domain transfer learning in building next-generation recommender systems, providing valuable insights for future advancements in the field.

人们对使用gnn来构建推荐系统越来越感兴趣,因为它们可以通过知识图嵌入来表示用户和项目之间的复杂关系。然而,大多数基于知识图的系统只关注评级或评论来建立关系。这阻碍了对图形数据中的结构和位置信息以及随时间变化的用户偏好的全面理解。为了解决这些问题,本文旨在提出一种先进的端到端图神经网络架构,该架构通过集成最先进的嵌入技术、知识图框架和迁移学习策略,显著增强了推荐系统的能力。结合位置编码和拓扑特征提取,提出的模型捕获复杂的用户-项目关系,并提供超越当前方法的鲁棒表示。预训练编码器促进知识转移,有效地弥合领域差距,提高预测精度。对已建立的基线模型的综合评估表明,我们的体系结构已经证明了增强的准确性、精确性和总体稳健性。这些结果强调了将知识图、复杂的嵌入策略和跨领域迁移学习结合起来构建下一代推荐系统的有效性,为该领域的未来发展提供了有价值的见解。
{"title":"Seasonality-Aware, Positional, and Topological-Guided GNN (SPT-GNN) for Movie Recommendation","authors":"Cevher Özden,&nbsp;Alper Özcan","doi":"10.1111/coin.70148","DOIUrl":"https://doi.org/10.1111/coin.70148","url":null,"abstract":"<div>\u0000 \u0000 <p>There has been an increasing interest in using GNNs to build recommender systems as they enable the representation of complex relationships between users and items through knowledge graph embeddings. However, most of the knowledge-graph-based systems focus only on ratings or reviews to build relationships. This prevents a comprehensive understanding of structural and positional information within graph data as well as user preferences that can change in time, as well. In order to address these issues, this paper aims to propose an advanced end-to-end Graph Neural Network architecture that significantly enhances recommendation system capabilities through the integration of state-of-the-art embedding techniques, knowledge graph frameworks, and transfer learning strategies. Incorporating positional encoding and topological feature extraction, the proposed model captures intricate user–item relationships and offers a robust representation that surpasses current approaches. A pretrained encoder facilitates knowledge transfer, effectively bridging domain gaps and amplifying prediction accuracy. Comprehensive evaluations against established baseline models reveal that our architecture has demonstrated enhanced accuracy, precision, and overall robustness. These results highlight the efficacy of combining knowledge graphs, sophisticated embedding strategies, and cross-domain transfer learning in building next-generation recommender systems, providing valuable insights for future advancements in the field.</p>\u0000 </div>","PeriodicalId":55228,"journal":{"name":"Computational Intelligence","volume":"41 5","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145272891","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comparative Analysis of Segment Anything Model and U-Net for Breast Tumor Detection in Ultrasound and Mammography Images 分段Anything模型与U-Net在超声与乳腺x线影像中乳腺肿瘤检测中的比较分析
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-11 DOI: 10.1111/coin.70145
Mohsen Ahmadi, Masoumeh Farhadi Nia, Sara Asgarian, Kasra Danesh, Elyas Irankhah, Ahmad Gholizadeh Lonbar, Abbas Sharifi

In this study, the main objective is to develop an algorithm capable of identifying and delineating tumor regions in breast ultrasound (BUS) and mammographic images. The technique employs two advanced deep learning architectures, namely U-Net and pretrained SAM, for tumor segmentation. The U-Net model is specifically designed for medical image segmentation and leverages its deep convolutional neural network framework to extract meaningful features from input images. On the other hand, the pretrained SAM architecture incorporates a mechanism to capture spatial dependencies and generate segmentation results. Evaluation is conducted on a diverse dataset containing annotated tumor regions in BUS and mammographic images, covering both benign and malignant tumors. This dataset enables a comprehensive assessment of the algorithm's performance across different tumor types. Results demonstrate that the U-Net model outperforms the pretrained SAM architecture in accurately identifying and segmenting tumor regions in both BUS and mammographic images. The U-Net exhibits superior performance in challenging cases involving irregular shapes, indistinct boundaries, and high tumor heterogeneity. In contrast, the pretrained SAM architecture exhibits limitations in accurately identifying tumor areas, particularly for malignant tumors and objects with weak boundaries or complex shapes. These findings highlight the importance of selecting appropriate deep learning architectures tailored for medical image segmentation. The U-Net model showcases its potential as a robust and accurate tool for tumor detection, while the pretrained SAM architecture suggests the need for further improvements to enhance segmentation performance.

在这项研究中,主要目的是开发一种能够识别和描绘乳腺超声(BUS)和乳房x线摄影图像中肿瘤区域的算法。该技术采用两种先进的深度学习架构,即U-Net和预训练的SAM,用于肿瘤分割。U-Net模型是专门为医学图像分割而设计的,并利用其深度卷积神经网络框架从输入图像中提取有意义的特征。另一方面,预训练的SAM架构结合了一种捕获空间依赖关系并生成分割结果的机制。评估是在不同的数据集上进行的,这些数据集包含了BUS和乳腺x线摄影图像中标注的肿瘤区域,包括良性和恶性肿瘤。该数据集能够全面评估该算法在不同肿瘤类型中的性能。结果表明,U-Net模型在BUS和乳房x线摄影图像中准确识别和分割肿瘤区域方面优于预训练的SAM架构。U-Net在不规则形状、边界不清、肿瘤异质性高的病例中表现优异。相比之下,预训练的SAM架构在准确识别肿瘤区域方面存在局限性,特别是对于恶性肿瘤和边界弱或形状复杂的物体。这些发现强调了为医学图像分割选择合适的深度学习架构的重要性。U-Net模型显示了其作为肿瘤检测的鲁棒性和准确性工具的潜力,而预训练的SAM架构表明需要进一步改进以提高分割性能。
{"title":"Comparative Analysis of Segment Anything Model and U-Net for Breast Tumor Detection in Ultrasound and Mammography Images","authors":"Mohsen Ahmadi,&nbsp;Masoumeh Farhadi Nia,&nbsp;Sara Asgarian,&nbsp;Kasra Danesh,&nbsp;Elyas Irankhah,&nbsp;Ahmad Gholizadeh Lonbar,&nbsp;Abbas Sharifi","doi":"10.1111/coin.70145","DOIUrl":"https://doi.org/10.1111/coin.70145","url":null,"abstract":"<div>\u0000 \u0000 <p>In this study, the main objective is to develop an algorithm capable of identifying and delineating tumor regions in breast ultrasound (BUS) and mammographic images. The technique employs two advanced deep learning architectures, namely U-Net and pretrained SAM, for tumor segmentation. The U-Net model is specifically designed for medical image segmentation and leverages its deep convolutional neural network framework to extract meaningful features from input images. On the other hand, the pretrained SAM architecture incorporates a mechanism to capture spatial dependencies and generate segmentation results. Evaluation is conducted on a diverse dataset containing annotated tumor regions in BUS and mammographic images, covering both benign and malignant tumors. This dataset enables a comprehensive assessment of the algorithm's performance across different tumor types. Results demonstrate that the U-Net model outperforms the pretrained SAM architecture in accurately identifying and segmenting tumor regions in both BUS and mammographic images. The U-Net exhibits superior performance in challenging cases involving irregular shapes, indistinct boundaries, and high tumor heterogeneity. In contrast, the pretrained SAM architecture exhibits limitations in accurately identifying tumor areas, particularly for malignant tumors and objects with weak boundaries or complex shapes. These findings highlight the importance of selecting appropriate deep learning architectures tailored for medical image segmentation. The U-Net model showcases its potential as a robust and accurate tool for tumor detection, while the pretrained SAM architecture suggests the need for further improvements to enhance segmentation performance.</p>\u0000 </div>","PeriodicalId":55228,"journal":{"name":"Computational Intelligence","volume":"41 5","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145272890","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Graph Neural Network-Based Online Collaborative Filtering Using Transductive Node Embeddings 基于换能化节点嵌入的图神经网络在线协同过滤
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-11 DOI: 10.1111/coin.70144
Gábor Szűcs, Richárd Kiss

The field of recommendation systems is a hot topic thanks to the increasing number of available digital products and services. In connection with this topic, the research of Graph Neural Network solutions has played a significant role in recent years. Research and development of an online recommendation system that also manages the challenges of a rapidly changing environment are important from a practical point of view as well. Our aim was to develop an approach that possesses scalable inference and adaptation and uses latent features. The main contribution of this paper is the development of a candidate generation process for online collaborative filtering on implicit feedback data that can scale to large user and item bases. We proposed multiple ways how embeddings can be obtained in a fast and scalable way, namely Lookup, Inductive neighbor aggregation, Neighbor aggregation with importance scores, and GraphSAGE-based Graph Neural Network (GraphSAGE+) method with continuous representation update for online learning. By combining these inductive and transductive methods for the embeddings, we developed a novel online Collaborative Filtering approach. We evaluated our approach on two e-commerce datasets and found that it outperformed traditional recommendation algorithms such as Matrix Factorization.

由于可用的数字产品和服务越来越多,推荐系统领域成为一个热门话题。针对这一课题,近年来图神经网络解决方案的研究发挥了重要作用。从实用的角度来看,研究和开发在线推荐系统也很重要,该系统还可以管理快速变化的环境中的挑战。我们的目标是开发一种具有可扩展推理和适应并使用潜在特征的方法。本文的主要贡献是开发了一个候选生成过程,用于隐式反馈数据的在线协同过滤,该过程可以扩展到大型用户和项目基础。我们提出了多种快速、可扩展地获取嵌入的方法,即查找、归纳邻居聚合、带重要分数的邻居聚合和基于GraphSAGE的连续表示更新的在线学习图神经网络(GraphSAGE+)方法。通过将这些归纳和转换的嵌入方法相结合,我们开发了一种新的在线协同过滤方法。我们在两个电子商务数据集上评估了我们的方法,发现它优于传统的推荐算法,如矩阵分解。
{"title":"Graph Neural Network-Based Online Collaborative Filtering Using Transductive Node Embeddings","authors":"Gábor Szűcs,&nbsp;Richárd Kiss","doi":"10.1111/coin.70144","DOIUrl":"https://doi.org/10.1111/coin.70144","url":null,"abstract":"<p>The field of recommendation systems is a hot topic thanks to the increasing number of available digital products and services. In connection with this topic, the research of Graph Neural Network solutions has played a significant role in recent years. Research and development of an online recommendation system that also manages the challenges of a rapidly changing environment are important from a practical point of view as well. Our aim was to develop an approach that possesses scalable inference and adaptation and uses latent features. The main contribution of this paper is the development of a candidate generation process for online collaborative filtering on implicit feedback data that can scale to large user and item bases. We proposed multiple ways how embeddings can be obtained in a fast and scalable way, namely Lookup, Inductive neighbor aggregation, Neighbor aggregation with importance scores, and GraphSAGE-based Graph Neural Network (GraphSAGE+) method with continuous representation update for online learning. By combining these inductive and transductive methods for the embeddings, we developed a novel online Collaborative Filtering approach. We evaluated our approach on two e-commerce datasets and found that it outperformed traditional recommendation algorithms such as Matrix Factorization.</p>","PeriodicalId":55228,"journal":{"name":"Computational Intelligence","volume":"41 5","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/coin.70144","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145272814","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Focused Segmentation in Biomedical Imaging via Attention Driven GAN-UNet 关注驱动GAN-UNet在生物医学成像中的焦点分割
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-02 DOI: 10.1111/coin.70128
Anamika Rangra, Chandan Kumar

Brain tumor segmentation is critical for diagnosis, treatment planning, and evaluation. However, existing methods such as U-Net, FCN, and Mask R-CNN often struggle with capturing fine-grained tumor boundaries, handling complex tumor heterogeneity, and maintaining high sensitivity across different tumor subregions. To overcome these challenges, this study proposes an Attention-Driven GAN-UNet framework that integrates U-Net with Generative Adversarial Networks (GANs) and a Channel-Spatial Attention Module (CSAM). This innovative approach enhances segmentation accuracy and focus mapping by directing the network's attention to clinically relevant regions. Trained on the BraTS 2020 dataset, our method surpasses traditional techniques, achieving a Dice Similarity Coefficient (DSC) of 0.99. The proposed framework visualizes intricate tumor morphologies, reduces false positives, and offers robust computational efficiency, making AttnGAN-UNet a promising tool for clinical brain tumor segmentation and analysis.

脑肿瘤的分割对诊断、治疗计划和评估至关重要。然而,现有的方法,如U-Net、FCN和Mask R-CNN,往往难以捕获细粒度的肿瘤边界,处理复杂的肿瘤异质性,并保持不同肿瘤亚区域的高灵敏度。为了克服这些挑战,本研究提出了一个注意力驱动的GAN-UNet框架,该框架将U-Net与生成对抗网络(gan)和通道空间注意力模块(CSAM)集成在一起。这种创新的方法通过将网络的注意力引导到临床相关区域,提高了分割的准确性和焦点映射。在BraTS 2020数据集上训练,我们的方法超越了传统技术,实现了0.99的骰子相似系数(DSC)。所提出的框架可以可视化复杂的肿瘤形态,减少假阳性,并提供强大的计算效率,使AttnGAN-UNet成为临床脑肿瘤分割和分析的有前途的工具。
{"title":"Focused Segmentation in Biomedical Imaging via Attention Driven GAN-UNet","authors":"Anamika Rangra,&nbsp;Chandan Kumar","doi":"10.1111/coin.70128","DOIUrl":"https://doi.org/10.1111/coin.70128","url":null,"abstract":"<div>\u0000 \u0000 <p>Brain tumor segmentation is critical for diagnosis, treatment planning, and evaluation. However, existing methods such as U-Net, FCN, and Mask R-CNN often struggle with capturing fine-grained tumor boundaries, handling complex tumor heterogeneity, and maintaining high sensitivity across different tumor subregions. To overcome these challenges, this study proposes an Attention-Driven GAN-UNet framework that integrates U-Net with Generative Adversarial Networks (GANs) and a Channel-Spatial Attention Module (CSAM). This innovative approach enhances segmentation accuracy and focus mapping by directing the network's attention to clinically relevant regions. Trained on the BraTS 2020 dataset, our method surpasses traditional techniques, achieving a Dice Similarity Coefficient (DSC) of 0.99. The proposed framework visualizes intricate tumor morphologies, reduces false positives, and offers robust computational efficiency, making AttnGAN-UNet a promising tool for clinical brain tumor segmentation and analysis.</p>\u0000 </div>","PeriodicalId":55228,"journal":{"name":"Computational Intelligence","volume":"41 5","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-10-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145223950","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient Right-Decoupled Composite Manifold Optimization for Visual Inertial Odometry 有效的右解耦复合流形视觉惯性里程计优化
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-02 DOI: 10.1111/coin.70127
Yangyang Ning
<div> <p>A composite manifold is defined as a concatenation of noninteracting manifolds, which may experience some loss of accuracy and consistency when propagating IMU dynamics based on Lie theory. However, from the perspective of ordinary differential equation modeling in dynamics, they demonstrate similar convergence rates and reduced computational complexity in iterative manifold optimization. In this context, this paper proposes a right decoupled composite manifold <span></span><math> <semantics> <mrow> <mfenced> <mrow> <mi>SO</mi> <mo>(</mo> <mn>3</mn> <mo>)</mo> <mo>,</mo> <msup> <mrow> <mi>ℝ</mi> </mrow> <mrow> <mn>3</mn> </mrow> </msup> <mo>,</mo> <msup> <mrow> <mi>ℝ</mi> </mrow> <mrow> <mn>3</mn> </mrow> </msup> </mrow> </mfenced> </mrow> <annotation>$$ leftlangle mathbf{SO}(3),{mathbb{R}}^3,{mathbb{R}}^3rightrangle $$</annotation> </semantics></math> for visual-inertial sliding-window iterative optimization compared with other manifolds including chained translation and rotation <span></span><math> <semantics> <mrow> <mfenced> <mrow> <mi>SO</mi> <mo>(</mo> <mn>3</mn> <mo>)</mo> <mo>×</mo> <msup> <mrow> <mi>ℝ</mi> </mrow> <mrow> <mn>3</mn> </mrow> </msup> <mo>,</mo> <msup> <mrow> <mi>ℝ</mi> </mrow> <mrow> <mn>3</mn> </mrow> </msup> </mrow> </mfenced> </mrow> <annotation>$$ leftlangle mathbf{SO}(3)times {mathbb{R}}^3,{mathbb{R}}^3rightrangle $$</annotation> </semantics></math>, special Euclidean group <span></span><math> <semantics> <mrow> <mfenced>
复合流形被定义为非相互作用流形的串联,在基于李氏理论传播IMU动力学时,可能会出现一些准确性和一致性的损失。然而,从动力学常微分方程建模的角度来看,它们在迭代流形优化中表现出相似的收敛速度和降低的计算复杂度。在此背景下,本文提出了一种右解耦复合流形SO (3),是,与其他流形(包括链式平移和旋转)相比,用于视觉惯性滑动窗口迭代优化的1 / 3 $$ leftlangle mathbf{SO}(3),{mathbb{R}}^3,{mathbb{R}}^3rightrangle $$所以(3)x,y3 $$ leftlangle mathbf{SO}(3)times {mathbb{R}}^3,{mathbb{R}}^3rightrangle $$,特殊欧几里德群SE (3);y3 $$ leftlangle mathbf{SE}(3),{mathbb{R}}^3rightrangle $$,和扩展位姿SE 2 (3) $$ {mathbf{SE}}_2(3) $$关于方向、位置和速度的估计。此外,通过半旋转扩展位姿SE 2 (3) $$ {mathbf{SE}}_2(3) $$传播惯性测量单元(IMU)动力学,以保持IMU预积分的精度。此外,为了增强鲁棒性,采用了一种鲁棒化的柯西损失函数。通过静态和更具挑战性的动态环境的仿真和实验,对该方法进行了精度、效率和鲁棒性评价。此外,还以解析形式给出了视觉重投影残差和IMU预积分残差所需的雅可比矩阵,并进行了数值验证。
{"title":"Efficient Right-Decoupled Composite Manifold Optimization for Visual Inertial Odometry","authors":"Yangyang Ning","doi":"10.1111/coin.70127","DOIUrl":"https://doi.org/10.1111/coin.70127","url":null,"abstract":"&lt;div&gt;\u0000 \u0000 &lt;p&gt;A composite manifold is defined as a concatenation of noninteracting manifolds, which may experience some loss of accuracy and consistency when propagating IMU dynamics based on Lie theory. However, from the perspective of ordinary differential equation modeling in dynamics, they demonstrate similar convergence rates and reduced computational complexity in iterative manifold optimization. In this context, this paper proposes a right decoupled composite manifold &lt;span&gt;&lt;/span&gt;&lt;math&gt;\u0000 &lt;semantics&gt;\u0000 &lt;mrow&gt;\u0000 &lt;mfenced&gt;\u0000 &lt;mrow&gt;\u0000 &lt;mi&gt;SO&lt;/mi&gt;\u0000 &lt;mo&gt;(&lt;/mo&gt;\u0000 &lt;mn&gt;3&lt;/mn&gt;\u0000 &lt;mo&gt;)&lt;/mo&gt;\u0000 &lt;mo&gt;,&lt;/mo&gt;\u0000 &lt;msup&gt;\u0000 &lt;mrow&gt;\u0000 &lt;mi&gt;ℝ&lt;/mi&gt;\u0000 &lt;/mrow&gt;\u0000 &lt;mrow&gt;\u0000 &lt;mn&gt;3&lt;/mn&gt;\u0000 &lt;/mrow&gt;\u0000 &lt;/msup&gt;\u0000 &lt;mo&gt;,&lt;/mo&gt;\u0000 &lt;msup&gt;\u0000 &lt;mrow&gt;\u0000 &lt;mi&gt;ℝ&lt;/mi&gt;\u0000 &lt;/mrow&gt;\u0000 &lt;mrow&gt;\u0000 &lt;mn&gt;3&lt;/mn&gt;\u0000 &lt;/mrow&gt;\u0000 &lt;/msup&gt;\u0000 &lt;/mrow&gt;\u0000 &lt;/mfenced&gt;\u0000 &lt;/mrow&gt;\u0000 &lt;annotation&gt;$$ leftlangle mathbf{SO}(3),{mathbb{R}}^3,{mathbb{R}}^3rightrangle $$&lt;/annotation&gt;\u0000 &lt;/semantics&gt;&lt;/math&gt; for visual-inertial sliding-window iterative optimization compared with other manifolds including chained translation and rotation &lt;span&gt;&lt;/span&gt;&lt;math&gt;\u0000 &lt;semantics&gt;\u0000 &lt;mrow&gt;\u0000 &lt;mfenced&gt;\u0000 &lt;mrow&gt;\u0000 &lt;mi&gt;SO&lt;/mi&gt;\u0000 &lt;mo&gt;(&lt;/mo&gt;\u0000 &lt;mn&gt;3&lt;/mn&gt;\u0000 &lt;mo&gt;)&lt;/mo&gt;\u0000 &lt;mo&gt;×&lt;/mo&gt;\u0000 &lt;msup&gt;\u0000 &lt;mrow&gt;\u0000 &lt;mi&gt;ℝ&lt;/mi&gt;\u0000 &lt;/mrow&gt;\u0000 &lt;mrow&gt;\u0000 &lt;mn&gt;3&lt;/mn&gt;\u0000 &lt;/mrow&gt;\u0000 &lt;/msup&gt;\u0000 &lt;mo&gt;,&lt;/mo&gt;\u0000 &lt;msup&gt;\u0000 &lt;mrow&gt;\u0000 &lt;mi&gt;ℝ&lt;/mi&gt;\u0000 &lt;/mrow&gt;\u0000 &lt;mrow&gt;\u0000 &lt;mn&gt;3&lt;/mn&gt;\u0000 &lt;/mrow&gt;\u0000 &lt;/msup&gt;\u0000 &lt;/mrow&gt;\u0000 &lt;/mfenced&gt;\u0000 &lt;/mrow&gt;\u0000 &lt;annotation&gt;$$ leftlangle mathbf{SO}(3)times {mathbb{R}}^3,{mathbb{R}}^3rightrangle $$&lt;/annotation&gt;\u0000 &lt;/semantics&gt;&lt;/math&gt;, special Euclidean group &lt;span&gt;&lt;/span&gt;&lt;math&gt;\u0000 &lt;semantics&gt;\u0000 &lt;mrow&gt;\u0000 &lt;mfenced&gt;\u0000 ","PeriodicalId":55228,"journal":{"name":"Computational Intelligence","volume":"41 5","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-10-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145223951","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computational Intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1