首页 > 最新文献

International Journal of Intelligent Computing and Cybernetics最新文献

英文 中文
X-News dataset for online news categorization 用于在线新闻分类的 X-News 数据集
IF 2.2 Q3 COMPUTER SCIENCE, CYBERNETICS Pub Date : 2024-08-13 DOI: 10.1108/ijicc-04-2024-0184
Samia Nawaz Yousafzai, Hooria Shahbaz, Armughan Ali, Amreen Qamar, Inzamam Mashood Nasir, Sara Tehsin, R. Damaševičius
PurposeThe objective is to develop a more effective model that simplifies and accelerates the news classification process using advanced text mining and deep learning (DL) techniques. A distributed framework utilizing Bidirectional Encoder Representations from Transformers (BERT) was developed to classify news headlines. This approach leverages various text mining and DL techniques on a distributed infrastructure, aiming to offer an alternative to traditional news classification methods.Design/methodology/approachThis study focuses on the classification of distinct types of news by analyzing tweets from various news channels. It addresses the limitations of using benchmark datasets for news classification, which often result in models that are impractical for real-world applications.FindingsThe framework’s effectiveness was evaluated on a newly proposed dataset and two additional benchmark datasets from the Kaggle repository, assessing the performance of each text mining and classification method across these datasets. The results of this study demonstrate that the proposed strategy significantly outperforms other approaches in terms of accuracy and execution time. This indicates that the distributed framework, coupled with the use of BERT for text analysis, provides a robust solution for analyzing large volumes of data efficiently. The findings also highlight the value of the newly released corpus for further research in news classification and emotion classification, suggesting its potential to facilitate advancements in these areas.Originality/valueThis research introduces an innovative distributed framework for news classification that addresses the shortcomings of models trained on benchmark datasets. By utilizing cutting-edge techniques and a novel dataset, the study offers significant improvements in accuracy and processing speed. The release of the corpus represents a valuable contribution to the field, enabling further exploration into news and emotion classification. This work sets a new standard for the analysis of news data, offering practical implications for the development of more effective and efficient news classification systems.
目的 目标是利用先进的文本挖掘和深度学习(DL)技术开发一种更有效的模型,以简化和加速新闻分类过程。我们开发了一个分布式框架,利用来自变换器的双向编码器表示(BERT)对新闻标题进行分类。这种方法在分布式基础架构上利用了各种文本挖掘和深度学习技术,旨在为传统的新闻分类方法提供一种替代方案。研究结果在新提出的数据集和来自 Kaggle 数据库的另外两个基准数据集上评估了该框架的有效性,评估了每种文本挖掘和分类方法在这些数据集上的性能。研究结果表明,所提出的策略在准确性和执行时间方面明显优于其他方法。这表明,分布式框架加上使用 BERT 进行文本分析,为高效分析海量数据提供了强大的解决方案。研究结果还凸显了新发布的语料库对于进一步研究新闻分类和情感分类的价值,表明它具有促进这些领域进步的潜力。 原创性/价值 这项研究为新闻分类引入了一个创新的分布式框架,解决了在基准数据集上训练的模型的缺点。通过利用前沿技术和新颖的数据集,该研究在准确性和处理速度方面都有显著提高。该语料库的发布是对该领域的宝贵贡献,有助于进一步探索新闻和情感分类。这项工作为新闻数据分析设定了新标准,为开发更有效、更高效的新闻分类系统提供了实际意义。
{"title":"X-News dataset for online news categorization","authors":"Samia Nawaz Yousafzai, Hooria Shahbaz, Armughan Ali, Amreen Qamar, Inzamam Mashood Nasir, Sara Tehsin, R. Damaševičius","doi":"10.1108/ijicc-04-2024-0184","DOIUrl":"https://doi.org/10.1108/ijicc-04-2024-0184","url":null,"abstract":"PurposeThe objective is to develop a more effective model that simplifies and accelerates the news classification process using advanced text mining and deep learning (DL) techniques. A distributed framework utilizing Bidirectional Encoder Representations from Transformers (BERT) was developed to classify news headlines. This approach leverages various text mining and DL techniques on a distributed infrastructure, aiming to offer an alternative to traditional news classification methods.Design/methodology/approachThis study focuses on the classification of distinct types of news by analyzing tweets from various news channels. It addresses the limitations of using benchmark datasets for news classification, which often result in models that are impractical for real-world applications.FindingsThe framework’s effectiveness was evaluated on a newly proposed dataset and two additional benchmark datasets from the Kaggle repository, assessing the performance of each text mining and classification method across these datasets. The results of this study demonstrate that the proposed strategy significantly outperforms other approaches in terms of accuracy and execution time. This indicates that the distributed framework, coupled with the use of BERT for text analysis, provides a robust solution for analyzing large volumes of data efficiently. The findings also highlight the value of the newly released corpus for further research in news classification and emotion classification, suggesting its potential to facilitate advancements in these areas.Originality/valueThis research introduces an innovative distributed framework for news classification that addresses the shortcomings of models trained on benchmark datasets. By utilizing cutting-edge techniques and a novel dataset, the study offers significant improvements in accuracy and processing speed. The release of the corpus represents a valuable contribution to the field, enabling further exploration into news and emotion classification. This work sets a new standard for the analysis of news data, offering practical implications for the development of more effective and efficient news classification systems.","PeriodicalId":45291,"journal":{"name":"International Journal of Intelligent Computing and Cybernetics","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141919573","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
X-News dataset for online news categorization 用于在线新闻分类的 X-News 数据集
IF 2.2 Q3 COMPUTER SCIENCE, CYBERNETICS Pub Date : 2024-08-13 DOI: 10.1108/ijicc-04-2024-0184
Samia Nawaz Yousafzai, Hooria Shahbaz, Armughan Ali, Amreen Qamar, Inzamam Mashood Nasir, Sara Tehsin, R. Damaševičius
PurposeThe objective is to develop a more effective model that simplifies and accelerates the news classification process using advanced text mining and deep learning (DL) techniques. A distributed framework utilizing Bidirectional Encoder Representations from Transformers (BERT) was developed to classify news headlines. This approach leverages various text mining and DL techniques on a distributed infrastructure, aiming to offer an alternative to traditional news classification methods.Design/methodology/approachThis study focuses on the classification of distinct types of news by analyzing tweets from various news channels. It addresses the limitations of using benchmark datasets for news classification, which often result in models that are impractical for real-world applications.FindingsThe framework’s effectiveness was evaluated on a newly proposed dataset and two additional benchmark datasets from the Kaggle repository, assessing the performance of each text mining and classification method across these datasets. The results of this study demonstrate that the proposed strategy significantly outperforms other approaches in terms of accuracy and execution time. This indicates that the distributed framework, coupled with the use of BERT for text analysis, provides a robust solution for analyzing large volumes of data efficiently. The findings also highlight the value of the newly released corpus for further research in news classification and emotion classification, suggesting its potential to facilitate advancements in these areas.Originality/valueThis research introduces an innovative distributed framework for news classification that addresses the shortcomings of models trained on benchmark datasets. By utilizing cutting-edge techniques and a novel dataset, the study offers significant improvements in accuracy and processing speed. The release of the corpus represents a valuable contribution to the field, enabling further exploration into news and emotion classification. This work sets a new standard for the analysis of news data, offering practical implications for the development of more effective and efficient news classification systems.
目的 目标是利用先进的文本挖掘和深度学习(DL)技术开发一种更有效的模型,以简化和加速新闻分类过程。我们开发了一个分布式框架,利用来自变换器的双向编码器表示(BERT)对新闻标题进行分类。这种方法在分布式基础架构上利用了各种文本挖掘和深度学习技术,旨在为传统的新闻分类方法提供一种替代方案。研究结果在新提出的数据集和来自 Kaggle 数据库的另外两个基准数据集上评估了该框架的有效性,评估了每种文本挖掘和分类方法在这些数据集上的性能。研究结果表明,所提出的策略在准确性和执行时间方面明显优于其他方法。这表明,分布式框架加上使用 BERT 进行文本分析,为高效分析海量数据提供了强大的解决方案。研究结果还凸显了新发布的语料库对于进一步研究新闻分类和情感分类的价值,表明它具有促进这些领域进步的潜力。 原创性/价值 这项研究为新闻分类引入了一个创新的分布式框架,解决了在基准数据集上训练的模型的缺点。通过利用前沿技术和新颖的数据集,该研究在准确性和处理速度方面都有显著提高。该语料库的发布是对该领域的宝贵贡献,有助于进一步探索新闻和情感分类。这项工作为新闻数据分析设定了新标准,为开发更有效、更高效的新闻分类系统提供了实际意义。
{"title":"X-News dataset for online news categorization","authors":"Samia Nawaz Yousafzai, Hooria Shahbaz, Armughan Ali, Amreen Qamar, Inzamam Mashood Nasir, Sara Tehsin, R. Damaševičius","doi":"10.1108/ijicc-04-2024-0184","DOIUrl":"https://doi.org/10.1108/ijicc-04-2024-0184","url":null,"abstract":"PurposeThe objective is to develop a more effective model that simplifies and accelerates the news classification process using advanced text mining and deep learning (DL) techniques. A distributed framework utilizing Bidirectional Encoder Representations from Transformers (BERT) was developed to classify news headlines. This approach leverages various text mining and DL techniques on a distributed infrastructure, aiming to offer an alternative to traditional news classification methods.Design/methodology/approachThis study focuses on the classification of distinct types of news by analyzing tweets from various news channels. It addresses the limitations of using benchmark datasets for news classification, which often result in models that are impractical for real-world applications.FindingsThe framework’s effectiveness was evaluated on a newly proposed dataset and two additional benchmark datasets from the Kaggle repository, assessing the performance of each text mining and classification method across these datasets. The results of this study demonstrate that the proposed strategy significantly outperforms other approaches in terms of accuracy and execution time. This indicates that the distributed framework, coupled with the use of BERT for text analysis, provides a robust solution for analyzing large volumes of data efficiently. The findings also highlight the value of the newly released corpus for further research in news classification and emotion classification, suggesting its potential to facilitate advancements in these areas.Originality/valueThis research introduces an innovative distributed framework for news classification that addresses the shortcomings of models trained on benchmark datasets. By utilizing cutting-edge techniques and a novel dataset, the study offers significant improvements in accuracy and processing speed. The release of the corpus represents a valuable contribution to the field, enabling further exploration into news and emotion classification. This work sets a new standard for the analysis of news data, offering practical implications for the development of more effective and efficient news classification systems.","PeriodicalId":45291,"journal":{"name":"International Journal of Intelligent Computing and Cybernetics","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141919130","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A novel ensemble causal feature selection approach with mutual information and group fusion strategy for multi-label data 针对多标签数据的互信息和组融合策略的新型集合因果特征选择方法
IF 2.2 Q3 COMPUTER SCIENCE, CYBERNETICS Pub Date : 2024-07-22 DOI: 10.1108/ijicc-04-2024-0144
Yifeng Zheng, Xianlong Zeng, Wenjie Zhang, Baoya Wei, Weishuo Ren, Depeng Qing
PurposeAs intelligent technology advances, practical applications often involve data with multiple labels. Therefore, multi-label feature selection methods have attracted much attention to extract valuable information. However, current methods tend to lack interpretability when evaluating the relationship between different types of variables without considering the potential causal relationship.Design/methodology/approachTo address the above problems, we propose an ensemble causal feature selection method based on mutual information and group fusion strategy (CMIFS) for multi-label data. First, the causal relationship between labels and features is analyzed by local causal structure learning, respectively, to obtain a causal feature set. Second, we eliminate false positive features from the obtained feature set using mutual information to improve the feature subset reliability. Eventually, we employ a group fusion strategy to fuse the obtained feature subsets from multiple data sub-space to enhance the stability of the results.FindingsExperimental comparisons are performed on six datasets to validate that our proposal can enhance the interpretation and robustness of the model compared with other methods in different metrics. Furthermore, the statistical analyses further validate the effectiveness of our approach.Originality/valueThe present study makes a noteworthy contribution to proposing a causal feature selection approach based on mutual information to obtain an approximate optimal feature subset for multi-label data. Additionally, our proposal adopts the group fusion strategy to guarantee the robustness of the obtained feature subset.
目的 随着智能技术的发展,实际应用中经常会涉及到带有多个标签的数据。因此,多标签特征选择方法在提取有价值信息方面备受关注。针对上述问题,我们提出了一种基于互信息和组融合策略(CMIFS)的多标签数据集合因果特征选择方法。首先,通过局部因果结构学习分别分析标签和特征之间的因果关系,得到因果特征集。其次,我们利用互信息从获得的特征集中剔除假阳性特征,以提高特征子集的可靠性。最后,我们采用分组融合策略,将从多个数据子空间获得的特征子集进行融合,以增强结果的稳定性。研究结果在六个数据集上进行了实验比较,验证了我们的建议与其他方法相比,能在不同指标上增强模型的解释性和鲁棒性。此外,统计分析进一步验证了我们方法的有效性。此外,我们的建议还采用了分组融合策略,以保证所获特征子集的鲁棒性。
{"title":"A novel ensemble causal feature selection approach with mutual information and group fusion strategy for multi-label data","authors":"Yifeng Zheng, Xianlong Zeng, Wenjie Zhang, Baoya Wei, Weishuo Ren, Depeng Qing","doi":"10.1108/ijicc-04-2024-0144","DOIUrl":"https://doi.org/10.1108/ijicc-04-2024-0144","url":null,"abstract":"PurposeAs intelligent technology advances, practical applications often involve data with multiple labels. Therefore, multi-label feature selection methods have attracted much attention to extract valuable information. However, current methods tend to lack interpretability when evaluating the relationship between different types of variables without considering the potential causal relationship.Design/methodology/approachTo address the above problems, we propose an ensemble causal feature selection method based on mutual information and group fusion strategy (CMIFS) for multi-label data. First, the causal relationship between labels and features is analyzed by local causal structure learning, respectively, to obtain a causal feature set. Second, we eliminate false positive features from the obtained feature set using mutual information to improve the feature subset reliability. Eventually, we employ a group fusion strategy to fuse the obtained feature subsets from multiple data sub-space to enhance the stability of the results.FindingsExperimental comparisons are performed on six datasets to validate that our proposal can enhance the interpretation and robustness of the model compared with other methods in different metrics. Furthermore, the statistical analyses further validate the effectiveness of our approach.Originality/valueThe present study makes a noteworthy contribution to proposing a causal feature selection approach based on mutual information to obtain an approximate optimal feature subset for multi-label data. Additionally, our proposal adopts the group fusion strategy to guarantee the robustness of the obtained feature subset.","PeriodicalId":45291,"journal":{"name":"International Journal of Intelligent Computing and Cybernetics","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-07-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141817315","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Contextualized dynamic meta embeddings based on Gated CNNs and self-attention for Arabic machine translation 基于门控 CNN 和自我关注的上下文动态元嵌入,用于阿拉伯语机器翻译
IF 2.2 Q3 COMPUTER SCIENCE, CYBERNETICS Pub Date : 2024-07-05 DOI: 10.1108/ijicc-03-2024-0106
Nouhaila Bensalah, H. Ayad, A. Adib, Abdelhamid Ibn El Farouk
PurposeThe paper aims to enhance Arabic machine translation (MT) by proposing novel approaches: (1) a dimensionality reduction technique for word embeddings tailored for Arabic text, optimizing efficiency while retaining semantic information; (2) a comprehensive comparison of meta-embedding techniques to improve translation quality; and (3) a method leveraging self-attention and Gated CNNs to capture token dependencies, including temporal and hierarchical features within sentences, and interactions between different embedding types. These approaches collectively aim to enhance translation quality by combining different embedding schemes and leveraging advanced modeling techniques.Design/methodology/approachRecent works on MT in general and Arabic MT in particular often pick one type of word embedding model. In this paper, we present a novel approach to enhance Arabic MT by addressing three key aspects. Firstly, we propose a new dimensionality reduction technique for word embeddings, specifically tailored for Arabic text. This technique optimizes the efficiency of embeddings while retaining their semantic information. Secondly, we conduct an extensive comparison of different meta-embedding techniques, exploring the combination of static and contextual embeddings. Through this analysis, we identify the most effective approach to improve translation quality. Lastly, we introduce a novel method that leverages self-attention and Gated convolutional neural networks (CNNs) to capture token dependencies, including temporal and hierarchical features within sentences, as well as interactions between different types of embeddings. Our experimental results demonstrate the effectiveness of our proposed approach in significantly enhancing Arabic MT performance. It outperforms baseline models with a BLEU score increase of 2 points and achieves superior results compared to state-of-the-art approaches, with an average improvement of 4.6 points across all evaluation metrics.FindingsThe proposed approaches significantly enhance Arabic MT performance. The dimensionality reduction technique improves the efficiency of word embeddings while preserving semantic information. Comprehensive comparison identifies effective meta-embedding techniques, with the contextualized dynamic meta-embeddings (CDME) model showcasing competitive results. Integration of Gated CNNs with the transformer model surpasses baseline performance, leveraging both architectures' strengths. Overall, these findings demonstrate substantial improvements in translation quality, with a BLEU score increase of 2 points and an average improvement of 4.6 points across all evaluation metrics, outperforming state-of-the-art approaches.Originality/valueThe paper’s originality lies in its departure from simply fine-tuning the transformer model for a specific task. Instead, it introduces modifications to the internal architecture of the transformer, integrating Gated CNNs to enhance translation performance. This
目的本文旨在通过提出以下新方法来提高阿拉伯语机器翻译(MT)的质量:(1)针对阿拉伯语文本量身定制的单词嵌入降维技术,在保留语义信息的同时优化效率;(2)全面比较元嵌入技术以提高翻译质量;以及(3)利用自我注意和门控 CNN 捕捉标记依赖性的方法,包括句子中的时间和层次特征,以及不同嵌入类型之间的交互。这些方法旨在通过结合不同的嵌入方案和利用先进的建模技术共同提高翻译质量。在本文中,我们针对三个关键方面提出了一种增强阿拉伯语 MT 的新方法。首先,我们提出了一种新的单词嵌入降维技术,专门针对阿拉伯语文本。该技术在保留嵌入词语义信息的同时,优化了嵌入词的效率。其次,我们对不同的元嵌入技术进行了广泛比较,探索了静态嵌入和上下文嵌入的结合。通过分析,我们确定了提高翻译质量的最有效方法。最后,我们介绍了一种新方法,该方法利用自我注意和门控卷积神经网络(CNN)来捕捉标记依赖性,包括句子中的时间和层次特征,以及不同类型嵌入之间的交互。实验结果表明,我们提出的方法在显著提高阿拉伯语 MT 性能方面非常有效。它优于基线模型,BLEU 分数提高了 2 分,并且与最先进的方法相比取得了更优异的结果,在所有评估指标上平均提高了 4.6 分。降维技术提高了词嵌入的效率,同时保留了语义信息。通过综合比较,确定了有效的元嵌入技术,其中语境化动态元嵌入(CDME)模型显示出极具竞争力的结果。门控 CNN 与转换器模型的集成超越了基线性能,充分利用了两种架构的优势。总体而言,这些研究结果表明翻译质量有了大幅提高,BLEU 分数提高了 2 分,所有评估指标平均提高了 4.6 分,超过了最先进的方法。 原创性/价值 本文的原创性在于它没有简单地针对特定任务对转换器模型进行微调。相反,它对转换器的内部架构进行了修改,整合了门控 CNN 以提高翻译性能。这种偏离传统微调方法的做法展示了一种全新的模型增强视角,为在不完全依赖现有架构的情况下提高翻译质量提供了独特的见解。降维的独创性在于为阿拉伯语文本量身定制的方法。虽然降维技术并不新鲜,但本文介绍了一种针对阿拉伯语词嵌入进行优化的特定方法。通过采用独立分量分析(ICA)和后处理方法,本文有效地降低了单词嵌入的维度,同时保留了语义信息,这在以前还没有过研究,尤其是在 MT 任务中。
{"title":"Contextualized dynamic meta embeddings based on Gated CNNs and self-attention for Arabic machine translation","authors":"Nouhaila Bensalah, H. Ayad, A. Adib, Abdelhamid Ibn El Farouk","doi":"10.1108/ijicc-03-2024-0106","DOIUrl":"https://doi.org/10.1108/ijicc-03-2024-0106","url":null,"abstract":"PurposeThe paper aims to enhance Arabic machine translation (MT) by proposing novel approaches: (1) a dimensionality reduction technique for word embeddings tailored for Arabic text, optimizing efficiency while retaining semantic information; (2) a comprehensive comparison of meta-embedding techniques to improve translation quality; and (3) a method leveraging self-attention and Gated CNNs to capture token dependencies, including temporal and hierarchical features within sentences, and interactions between different embedding types. These approaches collectively aim to enhance translation quality by combining different embedding schemes and leveraging advanced modeling techniques.Design/methodology/approachRecent works on MT in general and Arabic MT in particular often pick one type of word embedding model. In this paper, we present a novel approach to enhance Arabic MT by addressing three key aspects. Firstly, we propose a new dimensionality reduction technique for word embeddings, specifically tailored for Arabic text. This technique optimizes the efficiency of embeddings while retaining their semantic information. Secondly, we conduct an extensive comparison of different meta-embedding techniques, exploring the combination of static and contextual embeddings. Through this analysis, we identify the most effective approach to improve translation quality. Lastly, we introduce a novel method that leverages self-attention and Gated convolutional neural networks (CNNs) to capture token dependencies, including temporal and hierarchical features within sentences, as well as interactions between different types of embeddings. Our experimental results demonstrate the effectiveness of our proposed approach in significantly enhancing Arabic MT performance. It outperforms baseline models with a BLEU score increase of 2 points and achieves superior results compared to state-of-the-art approaches, with an average improvement of 4.6 points across all evaluation metrics.FindingsThe proposed approaches significantly enhance Arabic MT performance. The dimensionality reduction technique improves the efficiency of word embeddings while preserving semantic information. Comprehensive comparison identifies effective meta-embedding techniques, with the contextualized dynamic meta-embeddings (CDME) model showcasing competitive results. Integration of Gated CNNs with the transformer model surpasses baseline performance, leveraging both architectures' strengths. Overall, these findings demonstrate substantial improvements in translation quality, with a BLEU score increase of 2 points and an average improvement of 4.6 points across all evaluation metrics, outperforming state-of-the-art approaches.Originality/valueThe paper’s originality lies in its departure from simply fine-tuning the transformer model for a specific task. Instead, it introduces modifications to the internal architecture of the transformer, integrating Gated CNNs to enhance translation performance. This ","PeriodicalId":45291,"journal":{"name":"International Journal of Intelligent Computing and Cybernetics","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-07-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141674372","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dynamic community detection algorithm based on hyperbolic graph convolution 基于双曲图卷积的动态群落检测算法
IF 2.2 Q3 COMPUTER SCIENCE, CYBERNETICS Pub Date : 2024-07-04 DOI: 10.1108/ijicc-01-2024-0017
Weijiang Wu, Heping Tan, Yifeng Zheng
PurposeCommunity detection is a key factor in analyzing the structural features of complex networks. However, traditional dynamic community detection methods often fail to effectively solve the problems of deep network information loss and computational complexity in hyperbolic space. To address this challenge, a hyperbolic space-based dynamic graph neural network community detection model (HSDCDM) is proposed.Design/methodology/approachHSDCDM first projects the node features into the hyperbolic space and then utilizes the hyperbolic graph convolution module on the Poincaré and Lorentz models to realize feature fusion and information transfer. In addition, the parallel optimized temporal memory module ensures fast and accurate capture of time domain information over extended periods. Finally, the community clustering module divides the community structure by combining the node characteristics of the space domain and the time domain. To evaluate the performance of HSDCDM, experiments are conducted on both artificial and real datasets.FindingsExperimental results on complex networks demonstrate that HSDCDM significantly enhances the quality of community detection in hierarchical networks. It shows an average improvement of 7.29% in NMI and a 9.07% increase in ARI across datasets compared to traditional methods. For complex networks with non-Euclidean geometric structures, the HSDCDM model incorporating hyperbolic geometry can better handle the discontinuity of the metric space, provides a more compact embedding that preserves the data structure, and offers advantages over methods based on Euclidean geometry methods.Originality/valueThis model aggregates the potential information of nodes in space through manifold-preserving distribution mapping and hyperbolic graph topology modules. Moreover, it optimizes the Simple Recurrent Unit (SRU) on the hyperbolic space Lorentz model to effectively extract time series data in hyperbolic space, thereby enhancing computing efficiency by eliminating the reliance on tangent space.
目的群落检测是分析复杂网络结构特征的关键因素。然而,传统的动态群落检测方法往往无法有效解决双曲空间中的深度网络信息丢失和计算复杂性问题。为解决这一难题,本文提出了一种基于双曲空间的动态图神经网络群落检测模型(HSDCDM)。HSDCDM首先将节点特征投影到双曲空间,然后利用双曲图卷积模块对Poincaré和Lorentz模型实现特征融合和信息传递。此外,并行优化的时空存储模块可确保长时间快速、准确地捕捉时域信息。最后,群落聚类模块结合空间域和时间域的节点特征划分群落结构。对复杂网络的实验结果表明,HSDCDM 显著提高了分层网络中的群落检测质量。与传统方法相比,它在不同数据集上的 NMI 平均提高了 7.29%,ARI 平均提高了 9.07%。对于具有非欧几里得几何结构的复杂网络,包含双曲几何的 HSDCDM 模型能更好地处理度量空间的不连续性,提供了一种能保留数据结构的更紧凑的嵌入,与基于欧几里得几何的方法相比更具优势。此外,它还优化了双曲空间洛伦兹模型上的简单循环单元(SRU),以有效提取双曲空间中的时间序列数据,从而通过消除对切线空间的依赖来提高计算效率。
{"title":"Dynamic community detection algorithm based on hyperbolic graph convolution","authors":"Weijiang Wu, Heping Tan, Yifeng Zheng","doi":"10.1108/ijicc-01-2024-0017","DOIUrl":"https://doi.org/10.1108/ijicc-01-2024-0017","url":null,"abstract":"PurposeCommunity detection is a key factor in analyzing the structural features of complex networks. However, traditional dynamic community detection methods often fail to effectively solve the problems of deep network information loss and computational complexity in hyperbolic space. To address this challenge, a hyperbolic space-based dynamic graph neural network community detection model (HSDCDM) is proposed.Design/methodology/approachHSDCDM first projects the node features into the hyperbolic space and then utilizes the hyperbolic graph convolution module on the Poincaré and Lorentz models to realize feature fusion and information transfer. In addition, the parallel optimized temporal memory module ensures fast and accurate capture of time domain information over extended periods. Finally, the community clustering module divides the community structure by combining the node characteristics of the space domain and the time domain. To evaluate the performance of HSDCDM, experiments are conducted on both artificial and real datasets.FindingsExperimental results on complex networks demonstrate that HSDCDM significantly enhances the quality of community detection in hierarchical networks. It shows an average improvement of 7.29% in NMI and a 9.07% increase in ARI across datasets compared to traditional methods. For complex networks with non-Euclidean geometric structures, the HSDCDM model incorporating hyperbolic geometry can better handle the discontinuity of the metric space, provides a more compact embedding that preserves the data structure, and offers advantages over methods based on Euclidean geometry methods.Originality/valueThis model aggregates the potential information of nodes in space through manifold-preserving distribution mapping and hyperbolic graph topology modules. Moreover, it optimizes the Simple Recurrent Unit (SRU) on the hyperbolic space Lorentz model to effectively extract time series data in hyperbolic space, thereby enhancing computing efficiency by eliminating the reliance on tangent space.","PeriodicalId":45291,"journal":{"name":"International Journal of Intelligent Computing and Cybernetics","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141677447","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A novel ensemble artificial intelligence approach for coronary artery disease prediction 冠状动脉疾病预测的新型集合人工智能方法
IF 4.3 Q1 Computer Science Pub Date : 2024-06-06 DOI: 10.1108/ijicc-11-2023-0336
Ö. H. Namli, Seda Yanık, A. Erdoğan, Anke Schmeink
PurposeCoronary artery disease is one of the most common cardiovascular disorders in the world, and it can be deadly. Traditional diagnostic approaches are based on angiography, which is an interventional procedure having side effects such as contrast nephropathy or radio exposure as well as significant expenses. The purpose of this paper is to propose a novel artificial intelligence (AI) approach for the diagnosis of coronary artery disease as an effective alternative to traditional diagnostic methods.Design/methodology/approachIn this study, a novel ensemble AI approach based on optimization and classification is proposed. The proposed ensemble structure consists of three stages: feature selection, classification and combining. In the first stage, important features for each classification method are identified using the binary particle swarm optimization algorithm (BPSO). In the second stage, individual classification methods are used. In the final stage, the prediction results obtained from the individual methods are combined in an optimized way using the particle swarm optimization (PSO) algorithm to achieve better predictions.FindingsThe proposed method has been tested using an up-to-date real dataset collected at Basaksehir Çam and Sakura City Hospital. The data of disease prediction are unbalanced. Hence, the proposed ensemble approach improves majorly the F-measure and ROC area which are more prominent measures in case of unbalanced classification. The comparison shows that the proposed approach improves the F-measure and ROC area results of the individual classification methods around 14.5% in average and diagnoses with an accuracy rate of 96%.Originality/valueThis study presents a low-cost and low-risk AI-based approach for diagnosing heart disease compared to traditional diagnostic methods. Most of the existing research studies focus on base classification methods. In this study, we mainly investigate an effective ensemble method that uses optimization approaches for feature selection and combining stages for the medical diagnostic domain. Furthermore, the approaches in the literature are commonly tested on open-access dataset in heart disease diagnoses, whereas we apply our approach on a real and up-to-date dataset.
目的冠状动脉疾病是世界上最常见的心血管疾病之一,可致命。传统的诊断方法以血管造影术为基础,而血管造影术是一种介入性手术,具有造影剂肾病或放射线照射等副作用,而且费用高昂。本文旨在提出一种新型人工智能(AI)方法,用于诊断冠状动脉疾病,作为传统诊断方法的有效替代方案。所提出的集合结构包括三个阶段:特征选择、分类和组合。在第一阶段,使用二元粒子群优化算法(BPSO)确定每种分类方法的重要特征。在第二阶段,使用单个分类方法。在最后阶段,使用粒子群优化算法(PSO)以优化的方式合并从单个方法中获得的预测结果,以获得更好的预测结果。 研究结果使用在 Basaksehir Çam 和樱花市医院收集的最新真实数据集测试了所提出的方法。疾病预测数据是不平衡的。因此,所提出的集合方法主要提高了 F 值和 ROC 面积,而这两项指标在不平衡分类的情况下更为突出。比较结果表明,所提出的方法平均提高了单个分类方法的 F-measure 和 ROC 面积结果约 14.5%,诊断准确率高达 96%。现有的研究大多集中在基础分类方法上。在本研究中,我们主要研究了一种有效的集合方法,该方法在医疗诊断领域的特征选择和组合阶段使用了优化方法。此外,文献中的方法通常在心脏病诊断的开放数据集上进行测试,而我们的方法则应用于真实的最新数据集。
{"title":"A novel ensemble artificial intelligence approach for coronary artery disease prediction","authors":"Ö. H. Namli, Seda Yanık, A. Erdoğan, Anke Schmeink","doi":"10.1108/ijicc-11-2023-0336","DOIUrl":"https://doi.org/10.1108/ijicc-11-2023-0336","url":null,"abstract":"PurposeCoronary artery disease is one of the most common cardiovascular disorders in the world, and it can be deadly. Traditional diagnostic approaches are based on angiography, which is an interventional procedure having side effects such as contrast nephropathy or radio exposure as well as significant expenses. The purpose of this paper is to propose a novel artificial intelligence (AI) approach for the diagnosis of coronary artery disease as an effective alternative to traditional diagnostic methods.Design/methodology/approachIn this study, a novel ensemble AI approach based on optimization and classification is proposed. The proposed ensemble structure consists of three stages: feature selection, classification and combining. In the first stage, important features for each classification method are identified using the binary particle swarm optimization algorithm (BPSO). In the second stage, individual classification methods are used. In the final stage, the prediction results obtained from the individual methods are combined in an optimized way using the particle swarm optimization (PSO) algorithm to achieve better predictions.FindingsThe proposed method has been tested using an up-to-date real dataset collected at Basaksehir Çam and Sakura City Hospital. The data of disease prediction are unbalanced. Hence, the proposed ensemble approach improves majorly the F-measure and ROC area which are more prominent measures in case of unbalanced classification. The comparison shows that the proposed approach improves the F-measure and ROC area results of the individual classification methods around 14.5% in average and diagnoses with an accuracy rate of 96%.Originality/valueThis study presents a low-cost and low-risk AI-based approach for diagnosing heart disease compared to traditional diagnostic methods. Most of the existing research studies focus on base classification methods. In this study, we mainly investigate an effective ensemble method that uses optimization approaches for feature selection and combining stages for the medical diagnostic domain. Furthermore, the approaches in the literature are commonly tested on open-access dataset in heart disease diagnoses, whereas we apply our approach on a real and up-to-date dataset.","PeriodicalId":45291,"journal":{"name":"International Journal of Intelligent Computing and Cybernetics","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141380634","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
EYE-YOLO: a multi-spatial pyramid pooling and Focal-EIOU loss inspired tiny YOLOv7 for fundus eye disease detection EYE-YOLO:受微小 YOLOv7 启发,用于眼底疾病检测的多空间金字塔汇集和 Focal-EIOU 损失法
IF 4.3 Q1 Computer Science Pub Date : 2024-06-04 DOI: 10.1108/ijicc-02-2024-0077
Akhil Kumar, R. Dhanalakshmi
PurposeThe purpose of this work is to present an approach for autonomous detection of eye disease in fundus images. Furthermore, this work presents an improved variant of the Tiny YOLOv7 model developed specifically for eye disease detection. The model proposed in this work is a highly useful tool for the development of applications for autonomous detection of eye diseases in fundus images that can help and assist ophthalmologists.Design/methodology/approachThe approach adopted to carry out this work is twofold. Firstly, a richly annotated dataset consisting of eye disease classes, namely, cataract, glaucoma, retinal disease and normal eye, was created. Secondly, an improved variant of the Tiny YOLOv7 model was developed and proposed as EYE-YOLO. The proposed EYE-YOLO model has been developed by integrating multi-spatial pyramid pooling in the feature extraction network and Focal-EIOU loss in the detection network of the Tiny YOLOv7 model. Moreover, at run time, the mosaic augmentation strategy has been utilized with the proposed model to achieve benchmark results. Further, evaluations have been carried out for performance metrics, namely, precision, recall, F1 Score, average precision (AP) and mean average precision (mAP).FindingsThe proposed EYE-YOLO achieved 28% higher precision, 18% higher recall, 24% higher F1 Score and 30.81% higher mAP than the Tiny YOLOv7 model. Moreover, in terms of AP for each class of the employed dataset, it achieved 9.74% higher AP for cataract, 27.73% higher AP for glaucoma, 72.50% higher AP for retina disease and 13.26% higher AP for normal eye. In comparison to the state-of-the-art Tiny YOLOv5, Tiny YOLOv6 and Tiny YOLOv8 models, the proposed EYE-YOLO achieved 6–23.32% higher mAP.Originality/valueThis work addresses the problem of eye disease recognition as a bounding box regression and detection problem. Whereas, the work in the related research is largely based on eye disease classification. The other highlight of this work is to propose a richly annotated dataset for different eye diseases useful for training deep learning-based object detectors. The major highlight of this work lies in the proposal of an improved variant of the Tiny YOLOv7 model focusing on eye disease detection. The proposed modifications in the Tiny YOLOv7 aided the proposed model in achieving better results as compared to the state-of-the-art Tiny YOLOv8 and YOLOv8 Nano.
目的 本作品旨在介绍一种在眼底图像中自主检测眼疾的方法。此外,本作品还介绍了专门为眼疾检测开发的 Tiny YOLOv7 模型的改进变体。这项工作中提出的模型是一个非常有用的工具,可用于开发在眼底图像中自主检测眼部疾病的应用程序,从而帮助和协助眼科医生。首先,创建了一个包含丰富眼病类别注释的数据集,即白内障、青光眼、视网膜疾病和正常眼。其次,开发了 Tiny YOLOv7 模型的改进变体,并将其命名为 EYE-YOLO。所提出的 EYE-YOLO 模型是在 Tiny YOLOv7 模型的特征提取网络中集成了多空间金字塔池,在检测网络中集成了 Focal-EIOU 损失。此外,在运行时,还将马赛克增强策略与所提出的模型结合使用,以获得基准结果。此外,还对精确度、召回率、F1 分数、平均精确度 (AP) 和平均平均精确度 (mAP) 等性能指标进行了评估。此外,在所使用数据集的每个类别中,它在白内障方面的 AP 值提高了 9.74%,在青光眼方面的 AP 值提高了 27.73%,在视网膜疾病方面的 AP 值提高了 72.50%,在正常眼方面的 AP 值提高了 13.26%。与最先进的 Tiny YOLOv5、Tiny YOLOv6 和 Tiny YOLOv8 模型相比,所提出的 EYE-YOLO 的 mAP 高出 6-23.32%。而相关研究工作主要基于眼病分类。这项工作的另一个亮点是为不同的眼疾提出了丰富的注释数据集,有助于训练基于深度学习的物体检测器。这项工作的主要亮点在于提出了 Tiny YOLOv7 模型的改进变体,重点用于眼疾检测。与最先进的 Tiny YOLOv8 和 YOLOv8 Nano 相比,对 Tiny YOLOv7 提出的修改有助于该模型取得更好的结果。
{"title":"EYE-YOLO: a multi-spatial pyramid pooling and Focal-EIOU loss inspired tiny YOLOv7 for fundus eye disease detection","authors":"Akhil Kumar, R. Dhanalakshmi","doi":"10.1108/ijicc-02-2024-0077","DOIUrl":"https://doi.org/10.1108/ijicc-02-2024-0077","url":null,"abstract":"PurposeThe purpose of this work is to present an approach for autonomous detection of eye disease in fundus images. Furthermore, this work presents an improved variant of the Tiny YOLOv7 model developed specifically for eye disease detection. The model proposed in this work is a highly useful tool for the development of applications for autonomous detection of eye diseases in fundus images that can help and assist ophthalmologists.Design/methodology/approachThe approach adopted to carry out this work is twofold. Firstly, a richly annotated dataset consisting of eye disease classes, namely, cataract, glaucoma, retinal disease and normal eye, was created. Secondly, an improved variant of the Tiny YOLOv7 model was developed and proposed as EYE-YOLO. The proposed EYE-YOLO model has been developed by integrating multi-spatial pyramid pooling in the feature extraction network and Focal-EIOU loss in the detection network of the Tiny YOLOv7 model. Moreover, at run time, the mosaic augmentation strategy has been utilized with the proposed model to achieve benchmark results. Further, evaluations have been carried out for performance metrics, namely, precision, recall, F1 Score, average precision (AP) and mean average precision (mAP).FindingsThe proposed EYE-YOLO achieved 28% higher precision, 18% higher recall, 24% higher F1 Score and 30.81% higher mAP than the Tiny YOLOv7 model. Moreover, in terms of AP for each class of the employed dataset, it achieved 9.74% higher AP for cataract, 27.73% higher AP for glaucoma, 72.50% higher AP for retina disease and 13.26% higher AP for normal eye. In comparison to the state-of-the-art Tiny YOLOv5, Tiny YOLOv6 and Tiny YOLOv8 models, the proposed EYE-YOLO achieved 6–23.32% higher mAP.Originality/valueThis work addresses the problem of eye disease recognition as a bounding box regression and detection problem. Whereas, the work in the related research is largely based on eye disease classification. The other highlight of this work is to propose a richly annotated dataset for different eye diseases useful for training deep learning-based object detectors. The major highlight of this work lies in the proposal of an improved variant of the Tiny YOLOv7 model focusing on eye disease detection. The proposed modifications in the Tiny YOLOv7 aided the proposed model in achieving better results as compared to the state-of-the-art Tiny YOLOv8 and YOLOv8 Nano.","PeriodicalId":45291,"journal":{"name":"International Journal of Intelligent Computing and Cybernetics","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-06-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141266105","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Breast cancer pre-diagnosis based on incomplete picture fuzzy multi-granularity three-way decisions 基于不完整图像模糊多粒度三向决策的乳腺癌预诊断
IF 4.3 Q1 Computer Science Pub Date : 2024-06-04 DOI: 10.1108/ijicc-02-2024-0091
Haonan Hou, Chao Zhang, Fanghui Lu, Panna Lu
PurposeThree-way decision (3WD) and probabilistic rough sets (PRSs) are theoretical tools capable of simulating humans' multi-level and multi-perspective thinking modes in the field of decision-making. They are proposed to assist decision-makers in better managing incomplete or imprecise information under conditions of uncertainty or fuzziness. However, it is easy to cause decision losses and the personal thresholds of decision-makers cannot be taken into account. To solve this problem, this paper combines picture fuzzy (PF) multi-granularity (MG) with 3WD and establishes the notion of PF MG 3WD.Design/methodology/approachAn effective incomplete model based on PF MG 3WD is designed in this paper. First, the form of PF MG incomplete information systems (IISs) is established to reasonably record the uncertain information. On this basis, the PF conditional probability is established by using PF similarity relations, and the concept of adjustable PF MG PRSs is proposed by using the PF conditional probability to fuse data. Then, a comprehensive PF multi-attribute group decision-making (MAGDM) scheme is formed by the adjustable PF MG PRSs and the VlseKriterijumska Optimizacija I Kompromisno Resenje (VIKOR) method. Finally, an actual breast cancer data set is used to reveal the validity of the constructed method.FindingsThe experimental results confirm the effectiveness of PF MG 3WD in predicting breast cancer. Compared with existing models, PF MG 3WD has better robustness and generalization performance. This is mainly due to the incomplete PF MG 3WD proposed in this paper, which effectively reduces the influence of unreasonable outliers and threshold settings.Originality/valueThe model employs the VIKOR method for optimal granularity selections, which takes into account both group utility maximization and individual regret minimization, while incorporating decision-makers' subjective preferences as well. This ensures that the experiment maintains higher exclusion stability and reliability, enhancing the robustness of the decision results.
目的三向决策(3WD)和概率粗糙集(PRS)是决策领域能够模拟人类多层次、多角度思维模式的理论工具。它们的提出是为了帮助决策者在不确定性或模糊性条件下更好地管理不完整或不精确的信息。然而,它容易造成决策失误,而且无法考虑决策者的个人阈值。为解决这一问题,本文将图象模糊(PF)多粒度(MG)与 3WD 结合起来,建立了 PF MG 3WD 的概念。首先,建立了 PF MG 不完全信息系统(IIS)的形式,以合理记录不确定信息。在此基础上,利用 PF 相似性关系建立了 PF 条件概率,并利用 PF 条件概率融合数据,提出了可调整 PF MG PRS 的概念。然后,通过可调整 PF MG PRSs 和 VlseKriterijumska Optimizacija I Kompromisno Resenje(VIKOR)方法,形成了一个全面的 PF 多属性群体决策(MAGDM)方案。实验结果证实了 PF MG 3WD 预测乳腺癌的有效性。与现有模型相比,PF MG 3WD 具有更好的鲁棒性和泛化性能。原创性/价值该模型采用 VIKOR 方法进行最优粒度选择,既考虑了群体效用最大化和个体遗憾最小化,又结合了决策者的主观偏好。这确保了实验保持较高的排除稳定性和可靠性,增强了决策结果的稳健性。
{"title":"Breast cancer pre-diagnosis based on incomplete picture fuzzy multi-granularity three-way decisions","authors":"Haonan Hou, Chao Zhang, Fanghui Lu, Panna Lu","doi":"10.1108/ijicc-02-2024-0091","DOIUrl":"https://doi.org/10.1108/ijicc-02-2024-0091","url":null,"abstract":"PurposeThree-way decision (3WD) and probabilistic rough sets (PRSs) are theoretical tools capable of simulating humans' multi-level and multi-perspective thinking modes in the field of decision-making. They are proposed to assist decision-makers in better managing incomplete or imprecise information under conditions of uncertainty or fuzziness. However, it is easy to cause decision losses and the personal thresholds of decision-makers cannot be taken into account. To solve this problem, this paper combines picture fuzzy (PF) multi-granularity (MG) with 3WD and establishes the notion of PF MG 3WD.Design/methodology/approachAn effective incomplete model based on PF MG 3WD is designed in this paper. First, the form of PF MG incomplete information systems (IISs) is established to reasonably record the uncertain information. On this basis, the PF conditional probability is established by using PF similarity relations, and the concept of adjustable PF MG PRSs is proposed by using the PF conditional probability to fuse data. Then, a comprehensive PF multi-attribute group decision-making (MAGDM) scheme is formed by the adjustable PF MG PRSs and the VlseKriterijumska Optimizacija I Kompromisno Resenje (VIKOR) method. Finally, an actual breast cancer data set is used to reveal the validity of the constructed method.FindingsThe experimental results confirm the effectiveness of PF MG 3WD in predicting breast cancer. Compared with existing models, PF MG 3WD has better robustness and generalization performance. This is mainly due to the incomplete PF MG 3WD proposed in this paper, which effectively reduces the influence of unreasonable outliers and threshold settings.Originality/valueThe model employs the VIKOR method for optimal granularity selections, which takes into account both group utility maximization and individual regret minimization, while incorporating decision-makers' subjective preferences as well. This ensures that the experiment maintains higher exclusion stability and reliability, enhancing the robustness of the decision results.","PeriodicalId":45291,"journal":{"name":"International Journal of Intelligent Computing and Cybernetics","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-06-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141267660","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A hybrid approach for optimizing software defect prediction using a gray wolf optimization and multilayer perceptron 利用灰狼优化和多层感知器优化软件缺陷预测的混合方法
IF 4.3 Q1 Computer Science Pub Date : 2024-03-22 DOI: 10.1108/ijicc-11-2023-0385
Mohd. Mustaqeem, Suhel Mustajab, M.Aftab Alam
PurposeSoftware defect prediction (SDP) is a critical aspect of software quality assurance, aiming to identify and manage potential defects in software systems. In this paper, we have proposed a novel hybrid approach that combines Gray Wolf Optimization with Feature Selection (GWOFS) and multilayer perceptron (MLP) for SDP. The GWOFS-MLP hybrid model is designed to optimize feature selection, ultimately enhancing the accuracy and efficiency of SDP. Gray Wolf Optimization, inspired by the social hierarchy and hunting behavior of gray wolves, is employed to select a subset of relevant features from an extensive pool of potential predictors. This study investigates the key challenges that traditional SDP approaches encounter and proposes promising solutions to overcome time complexity and the curse of the dimensionality reduction problem.Design/methodology/approachThe integration of GWOFS and MLP results in a robust hybrid model that can adapt to diverse software datasets. This feature selection process harnesses the cooperative hunting behavior of wolves, allowing for the exploration of critical feature combinations. The selected features are then fed into an MLP, a powerful artificial neural network (ANN) known for its capability to learn intricate patterns within software metrics. MLP serves as the predictive engine, utilizing the curated feature set to model and classify software defects accurately.FindingsThe performance evaluation of the GWOFS-MLP hybrid model on a real-world software defect dataset demonstrates its effectiveness. The model achieves a remarkable training accuracy of 97.69% and a testing accuracy of 97.99%. Additionally, the receiver operating characteristic area under the curve (ROC-AUC) score of 0.89 highlights the model’s ability to discriminate between defective and defect-free software components.Originality/valueExperimental implementations using machine learning-based techniques with feature reduction are conducted to validate the proposed solutions. The goal is to enhance SDP’s accuracy, relevance and efficiency, ultimately improving software quality assurance processes. The confusion matrix further illustrates the model’s performance, with only a small number of false positives and false negatives.
目的软件缺陷预测(SDP)是软件质量保证的一个重要方面,旨在识别和管理软件系统中的潜在缺陷。本文提出了一种新颖的混合方法,将灰狼优化与特征选择(GWOFS)和多层感知器(MLP)结合起来用于 SDP。GWOFS-MLP 混合模型旨在优化特征选择,最终提高 SDP 的准确性和效率。灰狼优化的灵感来源于灰狼的社会等级制度和狩猎行为,用于从大量潜在预测因子中选择相关特征子集。本研究探讨了传统 SDP 方法所面临的主要挑战,并提出了有望克服时间复杂性和降维问题的解决方案。这一特征选择过程利用了狼的合作狩猎行为,允许探索关键特征组合。然后将选定的特征输入 MLP,这是一种功能强大的人工神经网络 (ANN),以能够学习软件度量中的复杂模式而著称。GWOFS-MLP 混合模型在实际软件缺陷数据集上的性能评估证明了它的有效性。该模型的训练准确率高达 97.69%,测试准确率高达 97.99%。此外,接收器工作特征曲线下面积(ROC-AUC)得分为 0.89,突出表明了该模型区分有缺陷和无缺陷软件组件的能力。目的是提高 SDP 的准确性、相关性和效率,最终改善软件质量保证流程。混淆矩阵进一步说明了该模型的性能,只有少量的误报和误判。
{"title":"A hybrid approach for optimizing software defect prediction using a gray wolf optimization and multilayer perceptron","authors":"Mohd. Mustaqeem, Suhel Mustajab, M.Aftab Alam","doi":"10.1108/ijicc-11-2023-0385","DOIUrl":"https://doi.org/10.1108/ijicc-11-2023-0385","url":null,"abstract":"PurposeSoftware defect prediction (SDP) is a critical aspect of software quality assurance, aiming to identify and manage potential defects in software systems. In this paper, we have proposed a novel hybrid approach that combines Gray Wolf Optimization with Feature Selection (GWOFS) and multilayer perceptron (MLP) for SDP. The GWOFS-MLP hybrid model is designed to optimize feature selection, ultimately enhancing the accuracy and efficiency of SDP. Gray Wolf Optimization, inspired by the social hierarchy and hunting behavior of gray wolves, is employed to select a subset of relevant features from an extensive pool of potential predictors. This study investigates the key challenges that traditional SDP approaches encounter and proposes promising solutions to overcome time complexity and the curse of the dimensionality reduction problem.Design/methodology/approachThe integration of GWOFS and MLP results in a robust hybrid model that can adapt to diverse software datasets. This feature selection process harnesses the cooperative hunting behavior of wolves, allowing for the exploration of critical feature combinations. The selected features are then fed into an MLP, a powerful artificial neural network (ANN) known for its capability to learn intricate patterns within software metrics. MLP serves as the predictive engine, utilizing the curated feature set to model and classify software defects accurately.FindingsThe performance evaluation of the GWOFS-MLP hybrid model on a real-world software defect dataset demonstrates its effectiveness. The model achieves a remarkable training accuracy of 97.69% and a testing accuracy of 97.99%. Additionally, the receiver operating characteristic area under the curve (ROC-AUC) score of 0.89 highlights the model’s ability to discriminate between defective and defect-free software components.Originality/valueExperimental implementations using machine learning-based techniques with feature reduction are conducted to validate the proposed solutions. The goal is to enhance SDP’s accuracy, relevance and efficiency, ultimately improving software quality assurance processes. The confusion matrix further illustrates the model’s performance, with only a small number of false positives and false negatives.","PeriodicalId":45291,"journal":{"name":"International Journal of Intelligent Computing and Cybernetics","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140213948","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Resource allocation in vehicular network based on sparrow search algorithm and hyper-graph in the presence of multiple cellular users 多蜂窝用户情况下基于麻雀搜索算法和超图的车载网络资源分配
IF 4.3 Q1 Computer Science Pub Date : 2024-01-25 DOI: 10.1108/ijicc-11-2023-0329
Lin Kang, Jie Wang, Junjie Chen, Di Yang
PurposeSince the performance of vehicular users and cellular users (CUE) in Vehicular networks is highly affected by the allocated resources to them. The purpose of this paper is to investigate the resource allocation for vehicular communications when multiple V2V links and a V2I link share spectrum with CUE in uplink communication under different Quality of Service (QoS).Design/methodology/approachAn optimization model to maximize the V2I capacity is established based on slowly varying large-scale fading channel information. Multiple V2V links are clustered based on sparrow search algorithm (SSA) to reduce interference. Then, a weighted tripartite graph is constructed by jointly optimizing the power of CUE, V2I and V2V clusters. Finally, spectrum resources are allocated based on a weighted 3D matching algorithm.FindingsThe performance of the proposed algorithm is tested. Simulation results show that the proposed algorithm can maximize the channel capacity of V2I while ensuring the reliability of V2V and the quality of service of CUE.Originality/valueThere is a lack of research on resource allocation algorithms of CUE, V2I and multiple V2V in different QoS. To solve the problem, one new resource allocation algorithm is proposed in this paper. Firstly, multiple V2V links are clustered using SSA to reduce interference. Secondly, the power allocation of CUE, V2I and V2V is jointly optimized. Finally, the weighted 3D matching algorithm is used to allocate spectrum resources.
目的由于车载网络中车辆用户和蜂窝用户(CUE)的性能受分配给他们的资源影响很大。本文旨在研究在不同服务质量(QoS)条件下,多个 V2V 链路和一个 V2I 链路在上行通信中与 CUE 共享频谱时的车载通信资源分配问题。基于麻雀搜索算法(SSA)对多个 V2V 链路进行聚类,以减少干扰。然后,通过联合优化 CUE、V2I 和 V2V 集群的功率,构建加权三方图。最后,根据加权三维匹配算法分配频谱资源。仿真结果表明,提出的算法可以最大限度地提高 V2I 的信道容量,同时确保 V2V 的可靠性和 CUE 的服务质量。为解决这一问题,本文提出了一种新的资源分配算法。首先,使用 SSA 对多个 V2V 链路进行聚类,以减少干扰。其次,联合优化 CUE、V2I 和 V2V 的功率分配。最后,使用加权三维匹配算法分配频谱资源。
{"title":"Resource allocation in vehicular network based on sparrow search algorithm and hyper-graph in the presence of multiple cellular users","authors":"Lin Kang, Jie Wang, Junjie Chen, Di Yang","doi":"10.1108/ijicc-11-2023-0329","DOIUrl":"https://doi.org/10.1108/ijicc-11-2023-0329","url":null,"abstract":"PurposeSince the performance of vehicular users and cellular users (CUE) in Vehicular networks is highly affected by the allocated resources to them. The purpose of this paper is to investigate the resource allocation for vehicular communications when multiple V2V links and a V2I link share spectrum with CUE in uplink communication under different Quality of Service (QoS).Design/methodology/approachAn optimization model to maximize the V2I capacity is established based on slowly varying large-scale fading channel information. Multiple V2V links are clustered based on sparrow search algorithm (SSA) to reduce interference. Then, a weighted tripartite graph is constructed by jointly optimizing the power of CUE, V2I and V2V clusters. Finally, spectrum resources are allocated based on a weighted 3D matching algorithm.FindingsThe performance of the proposed algorithm is tested. Simulation results show that the proposed algorithm can maximize the channel capacity of V2I while ensuring the reliability of V2V and the quality of service of CUE.Originality/valueThere is a lack of research on resource allocation algorithms of CUE, V2I and multiple V2V in different QoS. To solve the problem, one new resource allocation algorithm is proposed in this paper. Firstly, multiple V2V links are clustered using SSA to reduce interference. Secondly, the power allocation of CUE, V2I and V2V is jointly optimized. Finally, the weighted 3D matching algorithm is used to allocate spectrum resources.","PeriodicalId":45291,"journal":{"name":"International Journal of Intelligent Computing and Cybernetics","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-01-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139598370","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of Intelligent Computing and Cybernetics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1