Pub Date : 2024-10-18DOI: 10.1016/j.jksuci.2024.102219
Shijie Zeng , Yuefei Wang , Yukun Wen , Xi Yu , Binxiong Li , Zixu Wang
The Firefly Forest algorithm is a novel bio-inspired clustering method designed to address key challenges in traditional clustering techniques, such as the need to set a fixed number of neighbors, predefine cluster numbers, and rely on computationally intensive swarm iterative processes. The algorithm begins by using an adaptive neighbor estimation, refined to filter outliers, to determine the brightness of each firefly. This brightness guides the formation of firefly trees, which are then merged into cohesive firefly forests, completing the clustering process. This approach allows the algorithm to dynamically capture both local and global patterns, eliminate the need for predefined cluster numbers, and operate with low computational complexity. Experiments involving 14 established clustering algorithms across 19 diverse datasets, using 8 evaluative metrics, demonstrate the Firefly Forest algorithm’s superior accuracy and robustness. These results highlight its potential as a powerful tool for real-world clustering applications. Our code is available at: https://github.com/firesaku/FireflyForest.
{"title":"Firefly forest: A swarm iteration-free swarm intelligence clustering algorithm","authors":"Shijie Zeng , Yuefei Wang , Yukun Wen , Xi Yu , Binxiong Li , Zixu Wang","doi":"10.1016/j.jksuci.2024.102219","DOIUrl":"10.1016/j.jksuci.2024.102219","url":null,"abstract":"<div><div>The Firefly Forest algorithm is a novel bio-inspired clustering method designed to address key challenges in traditional clustering techniques, such as the need to set a fixed number of neighbors, predefine cluster numbers, and rely on computationally intensive swarm iterative processes. The algorithm begins by using an adaptive neighbor estimation, refined to filter outliers, to determine the brightness of each firefly. This brightness guides the formation of firefly trees, which are then merged into cohesive firefly forests, completing the clustering process. This approach allows the algorithm to dynamically capture both local and global patterns, eliminate the need for predefined cluster numbers, and operate with low computational complexity. Experiments involving 14 established clustering algorithms across 19 diverse datasets, using 8 evaluative metrics, demonstrate the Firefly Forest algorithm’s superior accuracy and robustness. These results highlight its potential as a powerful tool for real-world clustering applications. Our code is available at: https://github.com/firesaku/FireflyForest.</div></div>","PeriodicalId":48547,"journal":{"name":"Journal of King Saud University-Computer and Information Sciences","volume":"36 9","pages":"Article 102219"},"PeriodicalIF":5.2,"publicationDate":"2024-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142529769","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-09DOI: 10.1016/j.jksuci.2024.102205
Jiahao Li , Tao Luo, Baitao Zhang, Min Chen, Jie Zhou
With the development of data science, the challenge of high-dimensional data has become increasingly prevalent. High-dimensional data contains a significant amount of redundant information, which can adversely affect the performance and effectiveness of machine learning algorithms. Therefore, it is necessary to select the most relevant features from the raw data and perform feature selection on high-dimensional data. In this paper, a novel filter–wrapper feature selection method based on an improved multi-objective artificial bee colony algorithm (IMOABC) is proposed to address the feature selection problem in high-dimensional data. This method simultaneously considers three objectives: feature error rate, feature subset ratio, and distance, effectively improving the efficiency of obtaining the optimal feature subset on high-dimensional data. Additionally, a novel Fisher Score-based initialization strategy is introduced, significantly enhancing the quality of solutions. Furthermore, a new dynamic adaptive strategy is designed, effectively improving the algorithm’s convergence speed and enhancing its global search capability. Comparative experimental results on microarray cancer datasets demonstrate that the proposed method significantly improves classification accuracy and effectively reduces the size of the feature subset when compared to various traditional and state-of-the-art multi-objective feature selection algorithms. IMOABC improves the classification accuracy by 2.27% on average compared to various multi-objective feature selection methods, while reducing the number of selected features by 88.76% on average. Meanwhile, IMOABC shows an average improvement of 4.24% in classification accuracy across all datasets, with an average reduction of 76.73% in the number of selected features compared to various traditional methods.
{"title":"IMOABC: An efficient multi-objective filter–wrapper hybrid approach for high-dimensional feature selection","authors":"Jiahao Li , Tao Luo, Baitao Zhang, Min Chen, Jie Zhou","doi":"10.1016/j.jksuci.2024.102205","DOIUrl":"10.1016/j.jksuci.2024.102205","url":null,"abstract":"<div><div>With the development of data science, the challenge of high-dimensional data has become increasingly prevalent. High-dimensional data contains a significant amount of redundant information, which can adversely affect the performance and effectiveness of machine learning algorithms. Therefore, it is necessary to select the most relevant features from the raw data and perform feature selection on high-dimensional data. In this paper, a novel filter–wrapper feature selection method based on an improved multi-objective artificial bee colony algorithm (IMOABC) is proposed to address the feature selection problem in high-dimensional data. This method simultaneously considers three objectives: feature error rate, feature subset ratio, and distance, effectively improving the efficiency of obtaining the optimal feature subset on high-dimensional data. Additionally, a novel Fisher Score-based initialization strategy is introduced, significantly enhancing the quality of solutions. Furthermore, a new dynamic adaptive strategy is designed, effectively improving the algorithm’s convergence speed and enhancing its global search capability. Comparative experimental results on microarray cancer datasets demonstrate that the proposed method significantly improves classification accuracy and effectively reduces the size of the feature subset when compared to various traditional and state-of-the-art multi-objective feature selection algorithms. IMOABC improves the classification accuracy by 2.27% on average compared to various multi-objective feature selection methods, while reducing the number of selected features by 88.76% on average. Meanwhile, IMOABC shows an average improvement of 4.24% in classification accuracy across all datasets, with an average reduction of 76.73% in the number of selected features compared to various traditional methods.</div></div>","PeriodicalId":48547,"journal":{"name":"Journal of King Saud University-Computer and Information Sciences","volume":"36 9","pages":"Article 102205"},"PeriodicalIF":5.2,"publicationDate":"2024-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142432506","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-09DOI: 10.1016/j.jksuci.2024.102215
Abdul Wahid , Syed Zain Ul Abideen , Manzoor Ahmed , Wali Ullah Khan , Muhammad Sheraz , Teong Chee Chuah , Ying Loong Lee
The rapid development of next-generation wireless networks has intensified the need for robust security measures, particularly in environments susceptible to eavesdropping. Simultaneous transmitting and reflecting reconfigurable intelligent surfaces (STAR-RIS) have emerged as a transformative technology, offering full-space coverage by manipulating electromagnetic wave propagation. However, the inherent flexibility of STAR-RIS introduces new vulnerabilities, making secure communication a significant challenge. To overcome these challenges, we propose a deep reinforcement learning (DRL) based secure and efficient beamforming optimization strategy, leveraging the deep deterministic policy gradient (DDPG) algorithm. By framing the problem as a Markov decision process (MDP), our approach enables the DDPG algorithm to learn optimal strategies for beamforming and transmission and reflection coefficients (TARCs) configurations. This method is specifically designed to optimize phase-shift coefficients within the STAR-RIS environment, effectively managing the coupled phase shifts and complex interactions that are critical for enhancing physical layer security (PLS). Through extensive simulations, we demonstrate that our DRL-based strategy not only outperforms traditional optimization techniques but also achieves real-time adaptive optimization, significantly improving both confidentiality and network efficiency. This research addresses key gaps in secure wireless communication and sets a new standard for future advancements in intelligent, adaptable network technologies.
{"title":"Advanced security measures in coupled phase-shift STAR-RIS networks: A DRL approach","authors":"Abdul Wahid , Syed Zain Ul Abideen , Manzoor Ahmed , Wali Ullah Khan , Muhammad Sheraz , Teong Chee Chuah , Ying Loong Lee","doi":"10.1016/j.jksuci.2024.102215","DOIUrl":"10.1016/j.jksuci.2024.102215","url":null,"abstract":"<div><div>The rapid development of next-generation wireless networks has intensified the need for robust security measures, particularly in environments susceptible to eavesdropping. Simultaneous transmitting and reflecting reconfigurable intelligent surfaces (STAR-RIS) have emerged as a transformative technology, offering full-space coverage by manipulating electromagnetic wave propagation. However, the inherent flexibility of STAR-RIS introduces new vulnerabilities, making secure communication a significant challenge. To overcome these challenges, we propose a deep reinforcement learning (DRL) based secure and efficient beamforming optimization strategy, leveraging the deep deterministic policy gradient (DDPG) algorithm. By framing the problem as a Markov decision process (MDP), our approach enables the DDPG algorithm to learn optimal strategies for beamforming and transmission and reflection coefficients (TARCs) configurations. This method is specifically designed to optimize phase-shift coefficients within the STAR-RIS environment, effectively managing the coupled phase shifts and complex interactions that are critical for enhancing physical layer security (PLS). Through extensive simulations, we demonstrate that our DRL-based strategy not only outperforms traditional optimization techniques but also achieves real-time adaptive optimization, significantly improving both confidentiality and network efficiency. This research addresses key gaps in secure wireless communication and sets a new standard for future advancements in intelligent, adaptable network technologies.</div></div>","PeriodicalId":48547,"journal":{"name":"Journal of King Saud University-Computer and Information Sciences","volume":"36 9","pages":"Article 102215"},"PeriodicalIF":5.2,"publicationDate":"2024-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142432505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-05DOI: 10.1016/j.jksuci.2024.102208
Xiangwei Zheng, Dejian Su, Xuanchi Chen, Mingzhe Zhang
Probe-based confocal laser endomicroscopy (pCLE) is a significant diagnostic instrument and is frequently utilized to diagnose the severity of gastric intestinal metaplasia (GIM). The physicians must comprehensively analyze video clips recorded with pCLE from the gastric antrum, gastric body, and gastric angle area to determine the patient’s condition. However, due to the limitations of the pCLE’s microscopic imaging structure, the gastric areas detected cannot be identified and recorded in real time, which may poses a risk of missing potential areas of disease occurrence and is not conducive to the subsequent precise treatment of the lesion area. Therefore, this paper proposes an endoscopic video aided identification method for identifying gastric areas (EVIGA), which are utilized for determining the detected areas of pCLE in real-time. Firstly, the start time of the diagnosis clip is determined by real-time detecting the working states of pCLE. Then, the endoscopic video clip is truncated according to the correspondence between pCLE and endoscopic video in the time sequence for detecting gastric areas. In order to accurately identify pCLE detected gastric areas, a probe-based confocal laser endomicroscopy diagnosis area identification model (pCLEDAM) is constructed with an hourglass convolution designed for single-frame feature extraction and a temporal feature-sensitive extraction structure for spatial feature extraction. The extracted feature maps are unfolded and fed into the fully connected layer to classify the detected areas. To validate the proposed method, 67 clinical confocal laser endomicroscopy diagnosis cases are collected from a tertiary care hospital, and 500 video clips are finally reserved after audited for dataset construction. Experiments show that the accuracy of area identification on the test dataset achieves 96.0% and is much higher than other related algorithms, achieving the accurate identification of pCLE detected areas.
{"title":"Endoscopic video aided identification method for gastric area","authors":"Xiangwei Zheng, Dejian Su, Xuanchi Chen, Mingzhe Zhang","doi":"10.1016/j.jksuci.2024.102208","DOIUrl":"10.1016/j.jksuci.2024.102208","url":null,"abstract":"<div><div>Probe-based confocal laser endomicroscopy (pCLE) is a significant diagnostic instrument and is frequently utilized to diagnose the severity of gastric intestinal metaplasia (GIM). The physicians must comprehensively analyze video clips recorded with pCLE from the gastric antrum, gastric body, and gastric angle area to determine the patient’s condition. However, due to the limitations of the pCLE’s microscopic imaging structure, the gastric areas detected cannot be identified and recorded in real time, which may poses a risk of missing potential areas of disease occurrence and is not conducive to the subsequent precise treatment of the lesion area. Therefore, this paper proposes an endoscopic video aided identification method for identifying gastric areas (EVIGA), which are utilized for determining the detected areas of pCLE in real-time. Firstly, the start time of the diagnosis clip is determined by real-time detecting the working states of pCLE. Then, the endoscopic video clip is truncated according to the correspondence between pCLE and endoscopic video in the time sequence for detecting gastric areas. In order to accurately identify pCLE detected gastric areas, a probe-based confocal laser endomicroscopy diagnosis area identification model (pCLEDAM) is constructed with an hourglass convolution designed for single-frame feature extraction and a temporal feature-sensitive extraction structure for spatial feature extraction. The extracted feature maps are unfolded and fed into the fully connected layer to classify the detected areas. To validate the proposed method, 67 clinical confocal laser endomicroscopy diagnosis cases are collected from a tertiary care hospital, and 500 video clips are finally reserved after audited for dataset construction. Experiments show that the accuracy of area identification on the test dataset achieves 96.0% and is much higher than other related algorithms, achieving the accurate identification of pCLE detected areas.</div></div>","PeriodicalId":48547,"journal":{"name":"Journal of King Saud University-Computer and Information Sciences","volume":"36 9","pages":"Article 102208"},"PeriodicalIF":5.2,"publicationDate":"2024-10-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142424361","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-05DOI: 10.1016/j.jksuci.2024.102207
Vid Keršič, Sašo Karakatič, Muhamed Turkanović
Zero-knowledge proofs introduce a mechanism to prove that certain computations were performed without revealing any underlying information and are used commonly in blockchain-based decentralized apps (dapps). This cryptographic technique addresses trust issues prevalent in blockchain applications, and has now been adapted for machine learning (ML) services, known as Zero-Knowledge Machine Learning (ZKML). By leveraging the distributed nature of blockchains, this approach enhances the trustworthiness of ML deployments, and opens up new possibilities for privacy-preserving and robust ML applications within dapps. This paper provides a comprehensive overview of the ZKML process and its critical components for verifying ML services on-chain. Furthermore, this paper explores how blockchain technology and smart contracts can offer verifiable, trustless proof that a specific ML model has been used correctly to perform inference, all without relying on a single trusted entity. Additionally, the paper compares and reviews existing frameworks for implementing ZKML in dapps, serving as a reference point for researchers interested in this emerging field.
零知识证明引入了一种机制,用于证明某些计算是在不透露任何底层信息的情况下进行的,常用于基于区块链的去中心化应用程序(dapps)。这种加密技术解决了区块链应用中普遍存在的信任问题,现在已被用于机器学习(ML)服务,即零知识机器学习(ZKML)。通过利用区块链的分布式特性,这种方法提高了 ML 部署的可信度,并为 dapps 中保护隐私和稳健的 ML 应用开辟了新的可能性。本文全面概述了 ZKML 流程及其用于验证链上 ML 服务的关键组件。此外,本文还探讨了区块链技术和智能合约如何提供可验证的无信任证明,证明特定的 ML 模型已被正确用于执行推理,而无需依赖单一的可信实体。此外,本文还比较和回顾了在 dapp 中实施 ZKML 的现有框架,为对这一新兴领域感兴趣的研究人员提供了参考。
{"title":"On-chain zero-knowledge machine learning: An overview and comparison","authors":"Vid Keršič, Sašo Karakatič, Muhamed Turkanović","doi":"10.1016/j.jksuci.2024.102207","DOIUrl":"10.1016/j.jksuci.2024.102207","url":null,"abstract":"<div><div>Zero-knowledge proofs introduce a mechanism to prove that certain computations were performed without revealing any underlying information and are used commonly in blockchain-based decentralized apps (dapps). This cryptographic technique addresses trust issues prevalent in blockchain applications, and has now been adapted for machine learning (ML) services, known as Zero-Knowledge Machine Learning (ZKML). By leveraging the distributed nature of blockchains, this approach enhances the trustworthiness of ML deployments, and opens up new possibilities for privacy-preserving and robust ML applications within dapps. This paper provides a comprehensive overview of the ZKML process and its critical components for verifying ML services on-chain. Furthermore, this paper explores how blockchain technology and smart contracts can offer verifiable, trustless proof that a specific ML model has been used correctly to perform inference, all without relying on a single trusted entity. Additionally, the paper compares and reviews existing frameworks for implementing ZKML in dapps, serving as a reference point for researchers interested in this emerging field.</div></div>","PeriodicalId":48547,"journal":{"name":"Journal of King Saud University-Computer and Information Sciences","volume":"36 9","pages":"Article 102207"},"PeriodicalIF":5.2,"publicationDate":"2024-10-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142424359","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-05DOI: 10.1016/j.jksuci.2024.102206
Chaoran Wang , Mingyang Wang , Xianjie Wang , Yingchun Tan
Objectives:
Sequential recommendation aims to recommend items that are relevant to users’ interests based on their existing interaction sequences. Current models lack in capturing users’ latent intentions and do not sufficiently consider sequence information during the modeling of users and items. Additionally, noise in user interaction sequences can affect the model’s optimization process.
Methods:
This paper introduces an intent perceived sequential recommendation model (IPSRM). IPSRM employs the generalized expectation–maximization (EM) framework, alternating between learning sequence representations and optimizing the model to better capture the underlying intentions of user interactions. Specifically, IPSRM maps unlabeled behavioral sequences into frequency domain filtering and random Gaussian distribution space. These mappings reduce the impact of noise and improve the learning of user behavior representations. Through clustering process, IPSRM captures users’ potential interaction intentions and incorporates them as one of the supervisions into the contrastive self-supervised learning process to guide the optimization process.
Results:
Experimental results on four standard datasets demonstrate the superiority of IPSRM. Comparative experiments also verify that IPSRM exhibits strong robustness under cold start and noisy interaction conditions.
Conclusions:
Capturing latent user intentions, integrating intention-based supervision into model optimization, and mitigating noise in sequential modeling significantly enhance the performance of sequential recommendation systems.
{"title":"IPSRM: An intent perceived sequential recommendation model","authors":"Chaoran Wang , Mingyang Wang , Xianjie Wang , Yingchun Tan","doi":"10.1016/j.jksuci.2024.102206","DOIUrl":"10.1016/j.jksuci.2024.102206","url":null,"abstract":"<div><h3>Objectives:</h3><div>Sequential recommendation aims to recommend items that are relevant to users’ interests based on their existing interaction sequences. Current models lack in capturing users’ latent intentions and do not sufficiently consider sequence information during the modeling of users and items. Additionally, noise in user interaction sequences can affect the model’s optimization process.</div></div><div><h3>Methods:</h3><div>This paper introduces an intent perceived sequential recommendation model (IPSRM). IPSRM employs the generalized expectation–maximization (EM) framework, alternating between learning sequence representations and optimizing the model to better capture the underlying intentions of user interactions. Specifically, IPSRM maps unlabeled behavioral sequences into frequency domain filtering and random Gaussian distribution space. These mappings reduce the impact of noise and improve the learning of user behavior representations. Through clustering process, IPSRM captures users’ potential interaction intentions and incorporates them as one of the supervisions into the contrastive self-supervised learning process to guide the optimization process.</div></div><div><h3>Results:</h3><div>Experimental results on four standard datasets demonstrate the superiority of IPSRM. Comparative experiments also verify that IPSRM exhibits strong robustness under cold start and noisy interaction conditions.</div></div><div><h3>Conclusions:</h3><div>Capturing latent user intentions, integrating intention-based supervision into model optimization, and mitigating noise in sequential modeling significantly enhance the performance of sequential recommendation systems.</div></div>","PeriodicalId":48547,"journal":{"name":"Journal of King Saud University-Computer and Information Sciences","volume":"36 9","pages":"Article 102206"},"PeriodicalIF":5.2,"publicationDate":"2024-10-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142424360","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-05DOI: 10.1016/j.jksuci.2024.102204
Md Jahid Hasan , Wan Siti Halimatul Munirah Wan Ahmad , Mohammad Faizal Ahmad Fauzi , Jenny Tung Hiong Lee , See Yee Khor , Lai Meng Looi , Fazly Salleh Abas , Afzan Adam , Elaine Wan Ling Chan
Histopathology image segmentation and classification are essential for diagnosing and treating breast cancer. This study introduced a highly accurate segmentation and classification for histopathology images using a single architecture. We utilized the famous segmentation architectures, SegNet and U-Net, and modified the decoder to attach ResNet, VGG and DenseNet to perform classification tasks. These hybrid models are integrated with Stardist as the backbone, and implemented in a real-time pathologist workflow with a graphical user interface. These models were trained and tested offline using the ER-IHC-stained private and H&E-stained public datasets (MoNuSeg). For real-time evaluation, the proposed model was evaluated using PR-IHC-stained glass slides. It achieved the highest segmentation pixel-based F1-score of 0.902 and 0.903 for private and public datasets respectively, and a classification-based F1-score of 0.833 for private dataset. The experiment shows the robustness of our method where a model trained on ER-IHC dataset able to perform well on real-time microscopy of PR-IHC slides on both 20x and 40x magnification. This will help the pathologists with a quick decision-making process.
{"title":"Real-time segmentation and classification of whole-slide images for tumor biomarker scoring","authors":"Md Jahid Hasan , Wan Siti Halimatul Munirah Wan Ahmad , Mohammad Faizal Ahmad Fauzi , Jenny Tung Hiong Lee , See Yee Khor , Lai Meng Looi , Fazly Salleh Abas , Afzan Adam , Elaine Wan Ling Chan","doi":"10.1016/j.jksuci.2024.102204","DOIUrl":"10.1016/j.jksuci.2024.102204","url":null,"abstract":"<div><div>Histopathology image segmentation and classification are essential for diagnosing and treating breast cancer. This study introduced a highly accurate segmentation and classification for histopathology images using a single architecture. We utilized the famous segmentation architectures, SegNet and U-Net, and modified the decoder to attach ResNet, VGG and DenseNet to perform classification tasks. These hybrid models are integrated with Stardist as the backbone, and implemented in a real-time pathologist workflow with a graphical user interface. These models were trained and tested offline using the ER-IHC-stained private and H&E-stained public datasets (MoNuSeg). For real-time evaluation, the proposed model was evaluated using PR-IHC-stained glass slides. It achieved the highest segmentation pixel-based F1-score of 0.902 and 0.903 for private and public datasets respectively, and a classification-based F1-score of 0.833 for private dataset. The experiment shows the robustness of our method where a model trained on ER-IHC dataset able to perform well on real-time microscopy of PR-IHC slides on both 20x and 40x magnification. This will help the pathologists with a quick decision-making process.</div></div>","PeriodicalId":48547,"journal":{"name":"Journal of King Saud University-Computer and Information Sciences","volume":"36 9","pages":"Article 102204"},"PeriodicalIF":5.2,"publicationDate":"2024-10-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142529768","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-03DOI: 10.1016/j.jksuci.2024.102202
Yu Zhong, Bo Shen
Extracting structured information from unstructured text is crucial for knowledge management and utilization, which is the goal of document-level relation extraction. Existing graph-based methods face issues with information confusion and integration, limiting the reasoning capabilities of the model. To tackle this problem, a dual-stream dynamic graph structural network is proposed to model documents from various perspectives. Leveraging the richness of document information, a static document heterogeneous graph is constructed. A dynamic heterogeneous document graph is then induced based on this foundation to facilitate global information aggregation for entity representation learning. Additionally, the static document graph is decomposed into multi-level static semantic graphs, and multi-layer dynamic semantic graphs are further induced, explicitly segregating information from different levels. Information from different streams is effectively integrated via an information integrator. To mitigate the interference of noise during the reasoning process, a noise regularization mechanism is also designed. The experimental results on three extensively utilized publicly accessible datasets for document-level relation extraction demonstrate that our model achieves F1 scores of 62.56%, 71.1%, and 86.9% on the DocRED, CDR, and GDA datasets, respectively, significantly outperforming the baselines. Further analysis also demonstrates the effectiveness of the model in multi-entity scenarios.
从非结构化文本中提取结构化信息对于知识管理和利用至关重要,这也是文档级关系提取的目标。现有的基于图的方法面临着信息混淆和整合的问题,限制了模型的推理能力。为解决这一问题,我们提出了一种双流动态图结构网络,从不同角度对文档进行建模。利用丰富的文档信息,构建静态文档异构图。然后在此基础上诱导出动态异构文档图,以促进实体表征学习的全局信息聚合。此外,静态文档图被分解成多层次的静态语义图,并进一步诱导出多层次的动态语义图,明确分离来自不同层次的信息。来自不同信息流的信息通过信息集成器进行有效集成。为了减少推理过程中的噪声干扰,还设计了噪声正则化机制。在三个广泛使用的公开文档级关系提取数据集上的实验结果表明,我们的模型在 DocRED、CDR 和 GDA 数据集上的 F1 分数分别达到了 62.56%、71.1% 和 86.9%,明显优于基线模型。进一步的分析还证明了该模型在多实体场景中的有效性。
{"title":"Dual-stream dynamic graph structure network for document-level relation extraction","authors":"Yu Zhong, Bo Shen","doi":"10.1016/j.jksuci.2024.102202","DOIUrl":"10.1016/j.jksuci.2024.102202","url":null,"abstract":"<div><div>Extracting structured information from unstructured text is crucial for knowledge management and utilization, which is the goal of document-level relation extraction. Existing graph-based methods face issues with information confusion and integration, limiting the reasoning capabilities of the model. To tackle this problem, a dual-stream dynamic graph structural network is proposed to model documents from various perspectives. Leveraging the richness of document information, a static document heterogeneous graph is constructed. A dynamic heterogeneous document graph is then induced based on this foundation to facilitate global information aggregation for entity representation learning. Additionally, the static document graph is decomposed into multi-level static semantic graphs, and multi-layer dynamic semantic graphs are further induced, explicitly segregating information from different levels. Information from different streams is effectively integrated via an information integrator. To mitigate the interference of noise during the reasoning process, a noise regularization mechanism is also designed. The experimental results on three extensively utilized publicly accessible datasets for document-level relation extraction demonstrate that our model achieves F1 scores of 62.56%, 71.1%, and 86.9% on the DocRED, CDR, and GDA datasets, respectively, significantly outperforming the baselines. Further analysis also demonstrates the effectiveness of the model in multi-entity scenarios.</div></div>","PeriodicalId":48547,"journal":{"name":"Journal of King Saud University-Computer and Information Sciences","volume":"36 9","pages":"Article 102202"},"PeriodicalIF":5.2,"publicationDate":"2024-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142424440","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01DOI: 10.1016/j.jksuci.2024.102203
Yingqi Lu , Xiangsuo Fan , Jinfeng Wang , Shaojun Chen , Jie Meng
Accurate segmentation of lung nodules is crucial for the early detection of lung cancer and other pulmonary diseases. Traditional segmentation methods face several challenges, such as the overlap between nodules and surrounding anatomical structures like blood vessels and bronchi, as well as the variability in nodule size and shape, which complicates the segmentation algorithms. Existing methods often inadequately address these issues, highlighting the need for a more effective solution. To address these challenges, this paper proposes an improved multi-scale parallel fusion encoding network, ParaU-Net. ParaU-Net enhances the segmentation accuracy and model performance by optimizing the encoding process, improving feature extraction, preserving down-sampling information, and expanding the receptive field. Specifically, the multi-scale parallel fusion mechanism introduced in ParaU-Net better captures the fine features of nodules and reduces interference from other structures. Experiments conducted on the LIDC (The Lung Image Database Consortium) public dataset demonstrate the excellent performance of ParaU-Net in segmentation tasks, with results showing an IoU of 87.15%, Dice of 92.16%, F1-score of 92.24%, F2-score of 92.33%, and F0.5-score of 92.69%. These results significantly outperform other advanced segmentation methods, validating the effectiveness and accuracy of the proposed model in lung nodule CT image analysis. The code is available at https://github.com/XiaoBai-Lyq/ParaU-Net.
{"title":"ParaU-Net: An improved UNet parallel coding network for lung nodule segmentation","authors":"Yingqi Lu , Xiangsuo Fan , Jinfeng Wang , Shaojun Chen , Jie Meng","doi":"10.1016/j.jksuci.2024.102203","DOIUrl":"10.1016/j.jksuci.2024.102203","url":null,"abstract":"<div><div>Accurate segmentation of lung nodules is crucial for the early detection of lung cancer and other pulmonary diseases. Traditional segmentation methods face several challenges, such as the overlap between nodules and surrounding anatomical structures like blood vessels and bronchi, as well as the variability in nodule size and shape, which complicates the segmentation algorithms. Existing methods often inadequately address these issues, highlighting the need for a more effective solution. To address these challenges, this paper proposes an improved multi-scale parallel fusion encoding network, ParaU-Net. ParaU-Net enhances the segmentation accuracy and model performance by optimizing the encoding process, improving feature extraction, preserving down-sampling information, and expanding the receptive field. Specifically, the multi-scale parallel fusion mechanism introduced in ParaU-Net better captures the fine features of nodules and reduces interference from other structures. Experiments conducted on the LIDC (The Lung Image Database Consortium) public dataset demonstrate the excellent performance of ParaU-Net in segmentation tasks, with results showing an IoU of 87.15%, Dice of 92.16%, F1-score of 92.24%, F2-score of 92.33%, and F0.5-score of 92.69%. These results significantly outperform other advanced segmentation methods, validating the effectiveness and accuracy of the proposed model in lung nodule CT image analysis. The code is available at <span><span>https://github.com/XiaoBai-Lyq/ParaU-Net</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":48547,"journal":{"name":"Journal of King Saud University-Computer and Information Sciences","volume":"36 9","pages":"Article 102203"},"PeriodicalIF":5.2,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142424358","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01DOI: 10.1016/j.jksuci.2024.102200
Fan Wang , Xiaochen Yuan , Yue Liu , Chan-Tong Lam
Lung auscultation is essential for early lung condition detection. Categorizing adventitious lung sounds requires expert discrimination by medical specialists. This paper details the features of LungNeXt, a novel classification model specifically designed for lung sound analysis. Furthermore, we propose two auxiliary methods: RandClipMix (RCM) for data augmentation and Enhanced Mel-Spectrogram for Feature Extraction (EMFE). RCM addresses the issue of data imbalance by randomly mixing clips within the same category to create new adventitious lung sounds. EMFE augments specific frequency bands in spectrograms to highlight adventitious features. These contributions enable LungNeXt to achieve outstanding performance. LungNeXt optimally integrates an appropriate number of NeXtblocks, ensuring superior performance and a lightweight model architecture. The proposed RCM and EMFE methods, along with the LungNeXt classification network, have been evaluated on the SPRSound dataset. Experimental results revealed a commendable score of 0.5699 for the lung sound five-category task on SPRSound. Specifically, the LungNeXt model is characterized by its efficiency, with only 3.804M parameters and a computational complexity of 0.659G FLOPS. This lightweight and efficient model is particularly well-suited for applications in electronic stethoscope back-end processing equipment, providing efficient diagnostic advice to physicians and patients.
{"title":"LungNeXt: A novel lightweight network utilizing enhanced mel-spectrogram for lung sound classification","authors":"Fan Wang , Xiaochen Yuan , Yue Liu , Chan-Tong Lam","doi":"10.1016/j.jksuci.2024.102200","DOIUrl":"10.1016/j.jksuci.2024.102200","url":null,"abstract":"<div><div>Lung auscultation is essential for early lung condition detection. Categorizing adventitious lung sounds requires expert discrimination by medical specialists. This paper details the features of LungNeXt, a novel classification model specifically designed for lung sound analysis. Furthermore, we propose two auxiliary methods: RandClipMix (RCM) for data augmentation and Enhanced Mel-Spectrogram for Feature Extraction (EMFE). RCM addresses the issue of data imbalance by randomly mixing clips within the same category to create new adventitious lung sounds. EMFE augments specific frequency bands in spectrograms to highlight adventitious features. These contributions enable LungNeXt to achieve outstanding performance. LungNeXt optimally integrates an appropriate number of NeXtblocks, ensuring superior performance and a lightweight model architecture. The proposed RCM and EMFE methods, along with the LungNeXt classification network, have been evaluated on the SPRSound dataset. Experimental results revealed a commendable score of 0.5699 for the lung sound five-category task on SPRSound. Specifically, the LungNeXt model is characterized by its efficiency, with only 3.804M parameters and a computational complexity of 0.659G FLOPS. This lightweight and efficient model is particularly well-suited for applications in electronic stethoscope back-end processing equipment, providing efficient diagnostic advice to physicians and patients.</div></div>","PeriodicalId":48547,"journal":{"name":"Journal of King Saud University-Computer and Information Sciences","volume":"36 8","pages":"Article 102200"},"PeriodicalIF":5.2,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142358143","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}