Pub Date : 2026-02-05eCollection Date: 2026-01-01DOI: 10.3389/frai.2026.1702544
Jessie Bravo-Jaico, Oscar Serquén, Roger Alarcón, Juan Eduardo Suarez-Rivadeneira, Wilfredo Ruiz-Camacho, Freddy A Manayay
Tourism in biodiversity-rich regions was among the sectors most severely disrupted by the COVID-19 pandemic, which amplified existing socioeconomic vulnerabilities and placed cultural and natural heritage conservation at risk. In the Peruvian Amazon, Bagua Province illustrates this challenge, where experiential tourism is central to local livelihoods yet lacks adaptive management tools to support a sustainable recovery. To address this gap, this study introduces an integrated approach that combines artificial intelligence with biodiversity conservation through the application of multilayer perceptron (MLP) neural networks. By analyzing two decades of domestic visitor data (2003-2023), the research explores how predictive modeling can inform tourism governance in fragile ecosystems. Two scenarios were evaluated: one incorporating the complete dataset and another excluding the anomalous year 2020, heavily disrupted by the pandemic. The findings show that MLP models are capable of capturing visitor dynamics and forecasting demand fluctuations with notable accuracy. This predictive capacity allows for more adaptive planning of ecologically sensitive sites, such as the Tsunsuntsa Waterfall, where balancing visitor inflows with ecological thresholds is essential to preventing overtourism. Beyond technical accuracy, the study highlights the strategic potential of artificial intelligence as a governance tool that strengthens resilience in post-pandemic contexts, offering actionable insights for harmonizing socioeconomic recovery with biodiversity preservation. By positioning neural networks as vital instruments for sustainable destination management, this research contributes a reproducible model that can be adapted to other vulnerable regions worldwide. It underscores the value of integrating advanced computational methods into tourism governance frameworks, ultimately bridging technology and conservation to foster long-term sustainability.
{"title":"Artificial intelligence for biodiversity and tourism governance: predictive insights from multilayer perceptron models in Amazonia.","authors":"Jessie Bravo-Jaico, Oscar Serquén, Roger Alarcón, Juan Eduardo Suarez-Rivadeneira, Wilfredo Ruiz-Camacho, Freddy A Manayay","doi":"10.3389/frai.2026.1702544","DOIUrl":"https://doi.org/10.3389/frai.2026.1702544","url":null,"abstract":"<p><p>Tourism in biodiversity-rich regions was among the sectors most severely disrupted by the COVID-19 pandemic, which amplified existing socioeconomic vulnerabilities and placed cultural and natural heritage conservation at risk. In the Peruvian Amazon, Bagua Province illustrates this challenge, where experiential tourism is central to local livelihoods yet lacks adaptive management tools to support a sustainable recovery. To address this gap, this study introduces an integrated approach that combines artificial intelligence with biodiversity conservation through the application of multilayer perceptron (MLP) neural networks. By analyzing two decades of domestic visitor data (2003-2023), the research explores how predictive modeling can inform tourism governance in fragile ecosystems. Two scenarios were evaluated: one incorporating the complete dataset and another excluding the anomalous year 2020, heavily disrupted by the pandemic. The findings show that MLP models are capable of capturing visitor dynamics and forecasting demand fluctuations with notable accuracy. This predictive capacity allows for more adaptive planning of ecologically sensitive sites, such as the Tsunsuntsa Waterfall, where balancing visitor inflows with ecological thresholds is essential to preventing overtourism. Beyond technical accuracy, the study highlights the strategic potential of artificial intelligence as a governance tool that strengthens resilience in post-pandemic contexts, offering actionable insights for harmonizing socioeconomic recovery with biodiversity preservation. By positioning neural networks as vital instruments for sustainable destination management, this research contributes a reproducible model that can be adapted to other vulnerable regions worldwide. It underscores the value of integrating advanced computational methods into tourism governance frameworks, ultimately bridging technology and conservation to foster long-term sustainability.</p>","PeriodicalId":33315,"journal":{"name":"Frontiers in Artificial Intelligence","volume":"9 ","pages":"1702544"},"PeriodicalIF":4.7,"publicationDate":"2026-02-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12916584/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147272167","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
To enable rapid, accurate grading of tunnel surrounding rock during construction, we propose a real-time grading method that integrates image processing with lightweight deep learning. We developed an automated pipeline that combines image-processing techniques and machine-learning algorithms to extract and classify characteristic parameters of tunnel surrounding rock, enabling real-time monitoring and classification at the tunnel palm surface. The study demonstrates that: (1) Following the proposed image-acquisition standards for rock and tunnel palm surfaces, images are converted to grayscale, denoised, enhanced, and normalized, which facilitates efficient and accurate extraction of structural features and improves the precision of classification parameters; (2) An optimized lithology identification and classification model was built, and a rock-hardness, strength, and integrity sensing approach based on the ShuffleNetV2 convolutional neural network was introduced to achieve real-time surrounding-rock grading. On an engineering site, the method attains 85% accuracy for lithology classification, 75% for rock-mass integrity, and 80% for overall surrounding-rock grade, confirming its feasibility and practical value. These results offer theoretical insight and engineering utility for the scientific evaluation of tunnel surrounding-rock grade.
{"title":"Real-time grading method of tunnel surrounding rock based on image recognition.","authors":"Yihuan Xiao, Hao Yuan, Qingye Shi, Zemin Qiu, Liao Tang, Yihua Yu, Yabin Li, Yin Pan, Qinghua Xiao","doi":"10.3389/frai.2026.1766828","DOIUrl":"https://doi.org/10.3389/frai.2026.1766828","url":null,"abstract":"<p><p>To enable rapid, accurate grading of tunnel surrounding rock during construction, we propose a real-time grading method that integrates image processing with lightweight deep learning. We developed an automated pipeline that combines image-processing techniques and machine-learning algorithms to extract and classify characteristic parameters of tunnel surrounding rock, enabling real-time monitoring and classification at the tunnel palm surface. The study demonstrates that: (1) Following the proposed image-acquisition standards for rock and tunnel palm surfaces, images are converted to grayscale, denoised, enhanced, and normalized, which facilitates efficient and accurate extraction of structural features and improves the precision of classification parameters; (2) An optimized lithology identification and classification model was built, and a rock-hardness, strength, and integrity sensing approach based on the ShuffleNetV2 convolutional neural network was introduced to achieve real-time surrounding-rock grading. On an engineering site, the method attains 85% accuracy for lithology classification, 75% for rock-mass integrity, and 80% for overall surrounding-rock grade, confirming its feasibility and practical value. These results offer theoretical insight and engineering utility for the scientific evaluation of tunnel surrounding-rock grade.</p>","PeriodicalId":33315,"journal":{"name":"Frontiers in Artificial Intelligence","volume":"9 ","pages":"1766828"},"PeriodicalIF":4.7,"publicationDate":"2026-02-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12916587/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147272245","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Objective: Although imaging and paraspinal muscle parameters are linked to postoperative recurrent lumbar disc herniation (PRLDH), micro-level texture characteristics and their interactions remain underexplored. This study applied deep learning (DL)-radiomics to quantify the microstructural heterogeneity of responsible intervertebral discs and paraspinal muscles (L3-S1), and assessed a combined disc-muscle model for predicting PRLDH.
Method: Clinical and imaging data from 170 lumbar disc herniation (LDH) patients undergoing percutaneous transforaminal endoscopic surgery (Jan 2022-Dec 2024) were retrospectively analyzed. DL and radiomics features were extracted from intervertebral discs and paraspinal muscles. Feature selection via mutual information was followed by construction of a DL-radiomics Radscore model. Internal validation used leave-one-out, 10-fold cross-validation, and bootstrapping. Pfirrmann grading performance was compared with the disc Radscore, and potential disc-muscle interactions were explored using optimal cutoffs.
Results: Among 170 patients, 39 had postoperative recurrence. Disc Radscore included 2 DL and 3 radiomics features, while muscle Radscore comprised 2 DL and 5 radiomics features. The disc Radscore demonstrated good predictive ability (AUC 0.857, 95% CI 0.797-0.918) across validation methods (AUC 0.846-0.857). Muscle Radscore showed moderate performance (AUC 0.718, 95% CI 0.627-0.809). Pfirrmann grade poorly predicted recurrence (AUC 0.506, 95% CI 0.412-0.600). Combined disc-muscle analysis was less stable than disc Radscore alone.
Conclusion: DL-radiomics-derived intervertebral disc Radscore robustly predicts PRLDH. While combined disc-muscle assessment is less consistent, their interactions may inform postoperative risk stratification and management in LDH patients.
目的:尽管影像学和棘旁肌参数与术后复发性腰椎间盘突出症(PRLDH)有关,但微观层面的纹理特征及其相互作用仍未得到充分探讨。本研究应用深度学习(DL)-放射组学来量化责任椎间盘和棘旁肌(L3-S1)的微观结构异质性,并评估椎间盘-肌肉联合模型预测PRLDH。方法:回顾性分析2017年1月~ 2024年12月170例经皮经椎间孔内窥镜手术治疗的腰椎间盘突出症(LDH)的临床及影像学资料。从椎间盘和棘旁肌肉中提取DL和放射组学特征。通过互信息选择特征,然后构建DL-radiomics Radscore模型。内部验证使用留一、10倍交叉验证和引导。将Pfirrmann分级性能与椎间盘Radscore进行比较,并使用最佳截止点探索潜在的椎间盘-肌肉相互作用。结果:170例患者中39例术后复发。椎间盘Radscore包括2 DL和3个放射组学特征,肌肉Radscore包括2 DL和5个放射组学特征。在不同的验证方法(AUC 0.846-0.857)中,disc Radscore表现出良好的预测能力(AUC 0.857, 95% CI 0.797-0.918)。肌肉Radscore表现为中等表现(AUC 0.718, 95% CI 0.627-0.809)。Pfirrmann分级难以预测复发(AUC 0.506, 95% CI 0.412-0.600)。联合椎间盘-肌肉分析比单独椎间盘Radscore更不稳定。结论:dl放射组学衍生的椎间盘Radscore可靠地预测PRLDH。虽然椎间盘-肌肉联合评估不太一致,但它们的相互作用可能为LDH患者术后风险分层和管理提供信息。
{"title":"Deep learning-radiomics assessment of intervertebral disc and paraspinal muscle heterogeneity for predicting postoperative recurrent lumbar disc herniation.","authors":"Guangdong Zhang, Ziqian Zhu, Haiyan Zheng, Xindong Chang, Fanyi Zeng, Jianwei Cui, Ming Tang, Shiwu Yin","doi":"10.3389/frai.2026.1757269","DOIUrl":"https://doi.org/10.3389/frai.2026.1757269","url":null,"abstract":"<p><strong>Objective: </strong>Although imaging and paraspinal muscle parameters are linked to postoperative recurrent lumbar disc herniation (PRLDH), micro-level texture characteristics and their interactions remain underexplored. This study applied deep learning (DL)-radiomics to quantify the microstructural heterogeneity of responsible intervertebral discs and paraspinal muscles (L3-S1), and assessed a combined disc-muscle model for predicting PRLDH.</p><p><strong>Method: </strong>Clinical and imaging data from 170 lumbar disc herniation (LDH) patients undergoing percutaneous transforaminal endoscopic surgery (Jan 2022-Dec 2024) were retrospectively analyzed. DL and radiomics features were extracted from intervertebral discs and paraspinal muscles. Feature selection via mutual information was followed by construction of a DL-radiomics Radscore model. Internal validation used leave-one-out, 10-fold cross-validation, and bootstrapping. Pfirrmann grading performance was compared with the disc Radscore, and potential disc-muscle interactions were explored using optimal cutoffs.</p><p><strong>Results: </strong>Among 170 patients, 39 had postoperative recurrence. Disc Radscore included 2 DL and 3 radiomics features, while muscle Radscore comprised 2 DL and 5 radiomics features. The disc Radscore demonstrated good predictive ability (AUC 0.857, 95% CI 0.797-0.918) across validation methods (AUC 0.846-0.857). Muscle Radscore showed moderate performance (AUC 0.718, 95% CI 0.627-0.809). Pfirrmann grade poorly predicted recurrence (AUC 0.506, 95% CI 0.412-0.600). Combined disc-muscle analysis was less stable than disc Radscore alone.</p><p><strong>Conclusion: </strong>DL-radiomics-derived intervertebral disc Radscore robustly predicts PRLDH. While combined disc-muscle assessment is less consistent, their interactions may inform postoperative risk stratification and management in LDH patients.</p>","PeriodicalId":33315,"journal":{"name":"Frontiers in Artificial Intelligence","volume":"9 ","pages":"1757269"},"PeriodicalIF":4.7,"publicationDate":"2026-02-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12913562/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146228828","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-04eCollection Date: 2026-01-01DOI: 10.3389/frai.2026.1716063
J Jenifa Sharon, L Jani Anbarasi
Introduction: Accurate segmentation of kidney masses and structure is essential for medical application including diagnosis and treatment. This research proposed the dual track hybrid VHUCS-Net architecture which effectively highlights structural size-shape variants, boundaries and complex structural features in kidney disease.
Methods: Efficient segmentation is achieved by integrating the transformer enhanced U-Net model with the contrast optimized Protuberance Detection Network (PDN) model. The process begins with analysing kidney images using a standard U-Net combined with Vision Transformer attention and a High Resolution Network (HRNet) which capture global dependencies while preserving high resolution features resulting in accurate segmentation of the kidney region. Also, the masked kidney image undergoes processing through a contrast optimized PDN model with multi scale pooling, contrast enhancement, boundary refinement and explicit feature fusion to segment the mass region thereby enhancing mass localization improving border identification and enabling accurate abnormality detection. The resulting features are fused to provide a refined mass segmentation result that exactly identifies the location and structural abnormalities.
Results: The VHUCS-Net model was evaluated using the kidney segmentation dataset achieving an intersection over union score of 0.9441 and a dice coefficient of 0.9712 showing outstanding segmentation precision.
Discussion: These results indicate improved diagnostic efficiency and support clinical decision making by providing more accurate and interpretable segmentation outputs. Moreover, VHUCS-Net is validated with additional publicly available datasets with image mask correspondence, therefore proving the model effectiveness and generalizability across many segmentation tasks. The results highlight the capability of the proposed VHUCS-Net model to enhance diagnostic accuracy and assist clinical decision making through more detailed and interpretable segmentation outcomes.
{"title":"Advanced kidney mass segmentation using VHUCS-Net with protuberance detection network.","authors":"J Jenifa Sharon, L Jani Anbarasi","doi":"10.3389/frai.2026.1716063","DOIUrl":"https://doi.org/10.3389/frai.2026.1716063","url":null,"abstract":"<p><strong>Introduction: </strong>Accurate segmentation of kidney masses and structure is essential for medical application including diagnosis and treatment. This research proposed the dual track hybrid VHUCS-Net architecture which effectively highlights structural size-shape variants, boundaries and complex structural features in kidney disease.</p><p><strong>Methods: </strong>Efficient segmentation is achieved by integrating the transformer enhanced U-Net model with the contrast optimized Protuberance Detection Network (PDN) model. The process begins with analysing kidney images using a standard U-Net combined with Vision Transformer attention and a High Resolution Network (HRNet) which capture global dependencies while preserving high resolution features resulting in accurate segmentation of the kidney region. Also, the masked kidney image undergoes processing through a contrast optimized PDN model with multi scale pooling, contrast enhancement, boundary refinement and explicit feature fusion to segment the mass region thereby enhancing mass localization improving border identification and enabling accurate abnormality detection. The resulting features are fused to provide a refined mass segmentation result that exactly identifies the location and structural abnormalities.</p><p><strong>Results: </strong>The VHUCS-Net model was evaluated using the kidney segmentation dataset achieving an intersection over union score of 0.9441 and a dice coefficient of 0.9712 showing outstanding segmentation precision.</p><p><strong>Discussion: </strong>These results indicate improved diagnostic efficiency and support clinical decision making by providing more accurate and interpretable segmentation outputs. Moreover, VHUCS-Net is validated with additional publicly available datasets with image mask correspondence, therefore proving the model effectiveness and generalizability across many segmentation tasks. The results highlight the capability of the proposed VHUCS-Net model to enhance diagnostic accuracy and assist clinical decision making through more detailed and interpretable segmentation outcomes.</p>","PeriodicalId":33315,"journal":{"name":"Frontiers in Artificial Intelligence","volume":"9 ","pages":"1716063"},"PeriodicalIF":4.7,"publicationDate":"2026-02-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12913575/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146228912","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-04eCollection Date: 2026-01-01DOI: 10.3389/frai.2026.1737532
Fidelis Fidelis Alu, Sunkanmi Oluwadare
Artificial intelligence (AI) has shown promise in supporting clinical decision making, yet adoption in healthcare remains limited by concerns regarding transparency, verifiability, and accountability of AI-generated recommendations. In particular, generative and data-driven CDS systems often provide outputs without clearly exposing the evidentiary basis or reasoning process underlying their conclusions. This article presents a conceptual framework for auditable and source-verified AI-based clinical decision support, grounded in principles from evidence-based medicine, data provenance, and trustworthy AI. The proposed architecture integrates a curated medical knowledge base with explicit provenance metadata, a retrieval-augmented reasoning (RAG) engine that links generated recommendations to identifiable clinical guidelines and peer-reviewed sources, and a tamper-evident audit logging mechanism that records system inputs, retrieved evidence, and inference steps for retrospective review. This work does not introduce a new algorithm nor report a prototype implementation; rather, it synthesizes existing technical approaches into a coherent system design intended to improve traceability, clinician trust, and regulatory readiness. Key feasibility challenges are discussed, including knowledge-base governance and updating, citation fidelity in RAG architectures, bias propagation from underlying evidence, latency and usability trade-offs, privacy considerations, and alignment with emerging regulatory frameworks such as FDA Software as a Medical Device guidance and the European Union Artificial Intelligence Act. The article concludes by outlining a staged evaluation roadmap involving simulation studies and clinician-centered user research to guide future implementation and empirical validation.
{"title":"An auditable and source-verified framework for clinical AI decision support: integrating retrieval-augmented generation with data provenance.","authors":"Fidelis Fidelis Alu, Sunkanmi Oluwadare","doi":"10.3389/frai.2026.1737532","DOIUrl":"https://doi.org/10.3389/frai.2026.1737532","url":null,"abstract":"<p><p>Artificial intelligence (AI) has shown promise in supporting clinical decision making, yet adoption in healthcare remains limited by concerns regarding transparency, verifiability, and accountability of AI-generated recommendations. In particular, generative and data-driven CDS systems often provide outputs without clearly exposing the evidentiary basis or reasoning process underlying their conclusions. This article presents a conceptual framework for auditable and source-verified AI-based clinical decision support, grounded in principles from evidence-based medicine, data provenance, and trustworthy AI. The proposed architecture integrates a curated medical knowledge base with explicit provenance metadata, a retrieval-augmented reasoning (RAG) engine that links generated recommendations to identifiable clinical guidelines and peer-reviewed sources, and a tamper-evident audit logging mechanism that records system inputs, retrieved evidence, and inference steps for retrospective review. This work does not introduce a new algorithm nor report a prototype implementation; rather, it synthesizes existing technical approaches into a coherent system design intended to improve traceability, clinician trust, and regulatory readiness. Key feasibility challenges are discussed, including knowledge-base governance and updating, citation fidelity in RAG architectures, bias propagation from underlying evidence, latency and usability trade-offs, privacy considerations, and alignment with emerging regulatory frameworks such as FDA Software as a Medical Device guidance and the European Union Artificial Intelligence Act. The article concludes by outlining a staged evaluation roadmap involving simulation studies and clinician-centered user research to guide future implementation and empirical validation.</p>","PeriodicalId":33315,"journal":{"name":"Frontiers in Artificial Intelligence","volume":"9 ","pages":"1737532"},"PeriodicalIF":4.7,"publicationDate":"2026-02-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12913532/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146228901","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-04eCollection Date: 2026-01-01DOI: 10.3389/frai.2026.1718193
Zongrui Tian, Jiasheng Tian
Introduction: This article studies a new clustering-based federated learning algorithm that leverages Kullback-Leibler (KL) divergence to tackle heterogeneous data in wireless sensing environments.
Methods: Firstly, highdimensional heterogeneous data is subjected to principal component analysis to generate dimension-reduced representations, thereby reducing computational complexity. Secondly, the KL divergence distances between each pair of clients are calculated, followed by clustering according to the minimum threshold. The new KL divergence distance between the aggregated clients and others is taken as the average of the two. Finally, the federated learning training is conducted within each cluster to obtain a personalized model based on the classic wireless datasets.
Results and discussion: After the personalized models are tested, clients are reclustered and the models are updated-that is, a series of iterative operations, the optimal number of clusters and recognition accuracy are obtained. The test results show that the proposed algorithm based on KL divergence has higher recognition accuracy than several reported ones.
{"title":"A new clustered federated learning algorithm for heterogeneous data in high-precision wireless sensing.","authors":"Zongrui Tian, Jiasheng Tian","doi":"10.3389/frai.2026.1718193","DOIUrl":"https://doi.org/10.3389/frai.2026.1718193","url":null,"abstract":"<p><strong>Introduction: </strong>This article studies a new clustering-based federated learning algorithm that leverages Kullback-Leibler (KL) divergence to tackle heterogeneous data in wireless sensing environments.</p><p><strong>Methods: </strong>Firstly, highdimensional heterogeneous data is subjected to principal component analysis to generate dimension-reduced representations, thereby reducing computational complexity. Secondly, the KL divergence distances between each pair of clients are calculated, followed by clustering according to the minimum threshold. The new KL divergence distance between the aggregated clients and others is taken as the average of the two. Finally, the federated learning training is conducted within each cluster to obtain a personalized model based on the classic wireless datasets.</p><p><strong>Results and discussion: </strong>After the personalized models are tested, clients are reclustered and the models are updated-that is, a series of iterative operations, the optimal number of clusters and recognition accuracy are obtained. The test results show that the proposed algorithm based on KL divergence has higher recognition accuracy than several reported ones.</p>","PeriodicalId":33315,"journal":{"name":"Frontiers in Artificial Intelligence","volume":"9 ","pages":"1718193"},"PeriodicalIF":4.7,"publicationDate":"2026-02-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12913454/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146228903","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-04eCollection Date: 2026-01-01DOI: 10.3389/frai.2026.1666674
Piyush Kumar Soni, Radhakrishna Rambola
Implicit aspect detection aims to identify aspect categories that are not explicitly mentioned in text, but existing models struggle with four persistent challenges: aspect ambiguity, where multiple latent aspects are implied by the same expression, data imbalance and sparsity of implicit cues, contextual noise and syntactic variability in unstructured user reviews, and aspect drift, where the relevance of implicit cues changes across sentences or domains. To address these issues, this paper proposes the Transformer-Enhanced Graph Aspect Analyzer (TEGAA), a unified framework that tightly integrates dynamic expert routing, semantic representation refinement, and hierarchical graph reasoning. First, a Dynamic Expert Transformer (DET) equipped with a Dynamic Adaptive Expert Engine (DAEE) mitigates syntactic complexity and contextual noise by dynamically routing tokens to specialized expert sub-networks based on contextual and syntactic-semantic cues, enabling robust feature extraction for ambiguous implicit expressions. Second, Semantic Contrastive Learning (SCL) directly addresses data imbalance and weak implicit signals by enforcing semantic coherence among contextually related samples while increasing separability from irrelevant ones, thereby improving discriminability of sparse implicit aspect cues. Third, implicit aspect ambiguity and aspect drift are handled through a Graph-Enhanced Hierarchical Aspect Detector (GE-HAD), which models word- and sentence-level dependencies via context-aware graph attention. The incorporation of Attention Sinks prevents dominant but irrelevant tokens from overshadowing subtle implicit cues, while Pyramid Pooling aggregates multi-scale contextual information to stabilize aspect inference across varying linguistic scopes. Finally, an iterative feedback loop aligns graph-level reasoning with transformer-level expert routing, enabling adaptive refinement of aspect representations. Experiments on three benchmark datasets-Mobile Reviews, SemEval14, and Sentihood-demonstrate that TEGAA consistently outperforms state-of-the-art methods, achieving F1-scores above 0.88, precision above 0.89, recall above 0.87, accuracy exceeding 89%, and AUC values above 0.89. These results confirm TEGAA's effectiveness in resolving implicit aspect ambiguity, handling noisy and imbalanced data, and maintaining robust performance across domains.
{"title":"TEGAA: transformer-enhanced graph aspect analyzer with semantic contrastive learning for implicit aspect detection.","authors":"Piyush Kumar Soni, Radhakrishna Rambola","doi":"10.3389/frai.2026.1666674","DOIUrl":"https://doi.org/10.3389/frai.2026.1666674","url":null,"abstract":"<p><p>Implicit aspect detection aims to identify aspect categories that are not explicitly mentioned in text, but existing models struggle with four persistent challenges: aspect ambiguity, where multiple latent aspects are implied by the same expression, data imbalance and sparsity of implicit cues, contextual noise and syntactic variability in unstructured user reviews, and aspect drift, where the relevance of implicit cues changes across sentences or domains. To address these issues, this paper proposes the Transformer-Enhanced Graph Aspect Analyzer (TEGAA), a unified framework that tightly integrates dynamic expert routing, semantic representation refinement, and hierarchical graph reasoning. First, a Dynamic Expert Transformer (DET) equipped with a Dynamic Adaptive Expert Engine (DAEE) mitigates syntactic complexity and contextual noise by dynamically routing tokens to specialized expert sub-networks based on contextual and syntactic-semantic cues, enabling robust feature extraction for ambiguous implicit expressions. Second, Semantic Contrastive Learning (SCL) directly addresses data imbalance and weak implicit signals by enforcing semantic coherence among contextually related samples while increasing separability from irrelevant ones, thereby improving discriminability of sparse implicit aspect cues. Third, implicit aspect ambiguity and aspect drift are handled through a Graph-Enhanced Hierarchical Aspect Detector (GE-HAD), which models word- and sentence-level dependencies via context-aware graph attention. The incorporation of Attention Sinks prevents dominant but irrelevant tokens from overshadowing subtle implicit cues, while Pyramid Pooling aggregates multi-scale contextual information to stabilize aspect inference across varying linguistic scopes. Finally, an iterative feedback loop aligns graph-level reasoning with transformer-level expert routing, enabling adaptive refinement of aspect representations. Experiments on three benchmark datasets-Mobile Reviews, SemEval14, and Sentihood-demonstrate that TEGAA consistently outperforms state-of-the-art methods, achieving F1-scores above 0.88, precision above 0.89, recall above 0.87, accuracy exceeding 89%, and AUC values above 0.89. These results confirm TEGAA's effectiveness in resolving implicit aspect ambiguity, handling noisy and imbalanced data, and maintaining robust performance across domains.</p>","PeriodicalId":33315,"journal":{"name":"Frontiers in Artificial Intelligence","volume":"9 ","pages":"1666674"},"PeriodicalIF":4.7,"publicationDate":"2026-02-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12913412/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146228851","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-02eCollection Date: 2025-01-01DOI: 10.3389/frai.2025.1731633
K V Deepsahith, Basineni Shashank, Bangipavan Kumar, Sherly Alphonse, Brindha Subburaj, Girish Subramanian
Diabetic retinopathy (DR) detection can be performed through both deep retinal representations and vascular biomarkers. This proposed work suggests a multimodal framework that combines deep features with vascular descriptors in transformer fusion architecture. Fundus images are preprocessed using CLAHE, Canny edge detection, Top-hat transformation, and U-Net vessel segmentation. Then, the images are passed through a convolutional block attention module (CBAM)-fused enhanced MobileNetV3 backbone for deep spatial feature extraction. In parallel, the segmented vasculature is skeletonized to create a vascular graph, and the descriptors are computed using fractal dimension analysis (FDA), artery-to-vein ratio (AVR), and gray level co-occurrence matrix (GLCM) texture. A graph neural network (GNN) then generates a global topology-aware embedding using all that information. The different modalities are integrated using a transformer-based cross-modal fusion, where the feature vectors from MobileNet and GNN-based vascular embeddings interact using multi-head cross-attention. The fused representation is then given to a Softmax classifier for DR prediction. The model demonstrates superior performance compared to traditional deep learning baselines, achieving an accuracy of 93.8%, a precision of 92.1%, a recall of 92.8%, and an AUC-ROC of 0.96 for the DR prediction in the Messidor-2 dataset. The proposed approach also achieves above 98% accuracy for Eyepacs and APTOS 2019 datasets for DR detection. The findings demonstrate that the proposed system provides a reliable framework compared with the existing state-of-the-art methods.
{"title":"Graph-enhanced multimodal fusion of vascular biomarkers and deep features for diabetic retinopathy detection.","authors":"K V Deepsahith, Basineni Shashank, Bangipavan Kumar, Sherly Alphonse, Brindha Subburaj, Girish Subramanian","doi":"10.3389/frai.2025.1731633","DOIUrl":"https://doi.org/10.3389/frai.2025.1731633","url":null,"abstract":"<p><p>Diabetic retinopathy (DR) detection can be performed through both deep retinal representations and vascular biomarkers. This proposed work suggests a multimodal framework that combines deep features with vascular descriptors in transformer fusion architecture. Fundus images are preprocessed using CLAHE, Canny edge detection, Top-hat transformation, and U-Net vessel segmentation. Then, the images are passed through a convolutional block attention module (CBAM)-fused enhanced MobileNetV3 backbone for deep spatial feature extraction. In parallel, the segmented vasculature is skeletonized to create a vascular graph, and the descriptors are computed using fractal dimension analysis (FDA), artery-to-vein ratio (AVR), and gray level co-occurrence matrix (GLCM) texture. A graph neural network (GNN) then generates a global topology-aware embedding using all that information. The different modalities are integrated using a transformer-based cross-modal fusion, where the feature vectors from MobileNet and GNN-based vascular embeddings interact using multi-head cross-attention. The fused representation is then given to a Softmax classifier for DR prediction. The model demonstrates superior performance compared to traditional deep learning baselines, achieving an accuracy of 93.8%, a precision of 92.1%, a recall of 92.8%, and an AUC-ROC of 0.96 for the DR prediction in the Messidor-2 dataset. The proposed approach also achieves above 98% accuracy for Eyepacs and APTOS 2019 datasets for DR detection. The findings demonstrate that the proposed system provides a reliable framework compared with the existing state-of-the-art methods.</p>","PeriodicalId":33315,"journal":{"name":"Frontiers in Artificial Intelligence","volume":"8 ","pages":"1731633"},"PeriodicalIF":4.7,"publicationDate":"2026-02-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12907448/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146214290","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-02eCollection Date: 2026-01-01DOI: 10.3389/frai.2026.1748735
Hanlin Zhou, Huah Yong Chan
Introduction: Multi-agent/ensemble approaches can improve discrete-choice reasoning with large language models, but common orchestration methods are often non-deterministic, expensive, and difficult to reproduce. We propose ORCH, a deterministic multi-agent orchestrator that targets higher accuracy and better cost-performance via stable routing.
Methods: ORCH uses a pool of heterogeneous LLM agents and a deterministic routing mechanism based on exponential moving average (EMA) performance tracking. For each question, ORCH selects a small subset of agents, obtains candidate answers, and merges them through a controlled aggregation procedure. We evaluate ORCH on multiple discrete-choice benchmarks and compare against single-model baselines and non-routed ensemble strategies under consistent prompting and scoring.
Results: ORCH delivers consistent accuracy improvements over the best low-cost single model and provides additional gains over high-cost single-model baselines on several tasks, while reducing reliance on always-invoking expensive models. The deterministic routing and merge pipeline improves stability across runs.
Discussion: ORCH demonstrates that deterministic EMA-guided routing can offer a practical and reproducible orchestration strategy for discrete-choice reasoning. This framework can be extended to additional tasks, agent pools, and preference-aware routing policies in future work.
{"title":"ORCH: many analyses, one merge-a deterministic multi-agent orchestrator for discrete-choice reasoning with EMA-guided routing.","authors":"Hanlin Zhou, Huah Yong Chan","doi":"10.3389/frai.2026.1748735","DOIUrl":"https://doi.org/10.3389/frai.2026.1748735","url":null,"abstract":"<p><strong>Introduction: </strong>Multi-agent/ensemble approaches can improve discrete-choice reasoning with large language models, but common orchestration methods are often non-deterministic, expensive, and difficult to reproduce. We propose ORCH, a deterministic multi-agent orchestrator that targets higher accuracy and better cost-performance via stable routing.</p><p><strong>Methods: </strong>ORCH uses a pool of heterogeneous LLM agents and a deterministic routing mechanism based on exponential moving average (EMA) performance tracking. For each question, ORCH selects a small subset of agents, obtains candidate answers, and merges them through a controlled aggregation procedure. We evaluate ORCH on multiple discrete-choice benchmarks and compare against single-model baselines and non-routed ensemble strategies under consistent prompting and scoring.</p><p><strong>Results: </strong>ORCH delivers consistent accuracy improvements over the best low-cost single model and provides additional gains over high-cost single-model baselines on several tasks, while reducing reliance on always-invoking expensive models. The deterministic routing and merge pipeline improves stability across runs.</p><p><strong>Discussion: </strong>ORCH demonstrates that deterministic EMA-guided routing can offer a practical and reproducible orchestration strategy for discrete-choice reasoning. This framework can be extended to additional tasks, agent pools, and preference-aware routing policies in future work.</p>","PeriodicalId":33315,"journal":{"name":"Frontiers in Artificial Intelligence","volume":"9 ","pages":"1748735"},"PeriodicalIF":4.7,"publicationDate":"2026-02-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12907423/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146214259","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Fraud in Savings and Credit Cooperative Organizations (SACCOs) remains a major challenge that undermines financial inclusion and sustainability in developing countries. This study conducted a systematic literature review to examine both traditional and emerging fraud patterns and evaluate fraud detection methods with emphasis on artificial intelligence and machine learning applications. A comprehensive structured search across Web of Science, Scopus, and Google Scholar yielded 28 peer-reviewed studies published between 2015 and 2025 that met eligibility and quality criteria. The findings reveal that traditional fraud patterns such as member collusion, embezzlement, and asset misappropriation coexist with emerging digital fraud such as mobile payment fraud, phishing, card fraud, and cryptocurrency scams. While rule-based and audit-based detection remain ineffective, machine learning has demonstrated significant promise for real-time detection but faces challenges related to class imbalance, interpretability, and data privacy. The review identified a weak Information and Communication Technology (ICT) infrastructure, the absence of SACCO-specific fraud detection models, and hybrid frameworks. It concludes that hybrid models that integrate traditional audit methods with machine learning are recommended for SACCO-specific fraud detection frameworks. This study emphasizes the need for future research on explainable AI and privacy-preserving analytics to enhance fraud resilience in SACCOs.
储蓄和信用合作组织(SACCOs)中的欺诈行为仍然是破坏发展中国家金融包容性和可持续性的主要挑战。本研究对传统的和新兴的欺诈模式进行了系统的文献综述,并评估了欺诈检测方法,重点是人工智能和机器学习的应用。通过对Web of Science、Scopus和b谷歌Scholar的全面结构化搜索,得出了2015年至2025年间发表的28项同行评议研究,这些研究符合资格和质量标准。调查结果显示,传统的欺诈模式,如会员串通、贪污和资产挪用,与新兴的数字欺诈,如移动支付欺诈、网络钓鱼、信用卡欺诈和加密货币欺诈并存。虽然基于规则和基于审计的检测仍然无效,但机器学习在实时检测方面表现出了巨大的希望,但面临着与类别不平衡、可解释性和数据隐私相关的挑战。审查发现信息和通信技术(ICT)基础设施薄弱,缺乏针对sacco的欺诈检测模型,以及混合框架。它的结论是,建议将传统审计方法与机器学习相结合的混合模型用于sacco特定的欺诈检测框架。本研究强调未来需要对可解释的人工智能和隐私保护分析进行研究,以增强sacco的欺诈抵御能力。
{"title":"A systematic review and future directions for AI-driven detection of fraud patterns in SACCO transactions.","authors":"Dalton Ampumuza, Calorine Katushabe, Micheal Tamale","doi":"10.3389/frai.2025.1690482","DOIUrl":"10.3389/frai.2025.1690482","url":null,"abstract":"<p><p>Fraud in Savings and Credit Cooperative Organizations (SACCOs) remains a major challenge that undermines financial inclusion and sustainability in developing countries. This study conducted a systematic literature review to examine both traditional and emerging fraud patterns and evaluate fraud detection methods with emphasis on artificial intelligence and machine learning applications. A comprehensive structured search across Web of Science, Scopus, and Google Scholar yielded 28 peer-reviewed studies published between 2015 and 2025 that met eligibility and quality criteria. The findings reveal that traditional fraud patterns such as member collusion, embezzlement, and asset misappropriation coexist with emerging digital fraud such as mobile payment fraud, phishing, card fraud, and cryptocurrency scams. While rule-based and audit-based detection remain ineffective, machine learning has demonstrated significant promise for real-time detection but faces challenges related to class imbalance, interpretability, and data privacy. The review identified a weak Information and Communication Technology (ICT) infrastructure, the absence of SACCO-specific fraud detection models, and hybrid frameworks. It concludes that hybrid models that integrate traditional audit methods with machine learning are recommended for SACCO-specific fraud detection frameworks. This study emphasizes the need for future research on explainable AI and privacy-preserving analytics to enhance fraud resilience in SACCOs.</p>","PeriodicalId":33315,"journal":{"name":"Frontiers in Artificial Intelligence","volume":"8 ","pages":"1690482"},"PeriodicalIF":4.7,"publicationDate":"2026-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12901501/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146202961","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}