Pub Date : 2025-10-23DOI: 10.1186/s42492-025-00206-w
Alexander Somerville, Keith Joiner, Timothy Lynar, Graham Wild
The use of extended reality (XR) spectrum technologies as substitutes to augment traditional simulators in pilot flight training has received significant interest in recent times. A systematic review was conducted to evaluate the efficacy of XR technologies for this purpose and better understand the motivating factors for this use. The systematic review followed the QUOROM framework (adapted for educational studies), screening 1237 candidate articles to 67 eligible for thematic analysis, with 5 of these also meeting meta-analysis criteria. Existing literature emphasizes the benefits of these technologies, particularly as a result of immersiveness and spatial awareness, enabling the application of more modern educational theories. Although the existing literature is concerned with much of the industry, there is a specific focus on general aviation and the more ab initio skills of flight. The results of the meta-analysis indicate improvements in pilot performance, with an overall meta-analytic effect size estimate of 0.884 (z = 2.248, P = 0.025), which is positive, statistically significant, and moderately strong. The findings of this review indicate support for the use and intention for the use of XR in pilot flight training simulators. However, multiple serious research gaps exist, such as the potential higher occurrence of simulator sickness and cybersickness, and a lack of robust research trials that examine transfer of training across the full pilot skill set and curricular contexts. This novel systematic review and meta-analysis represent a significant attempt to shape and direct better research to help to direct flourishing technological XR development in a time of increasing pilot shortages and aviation growth.
近年来,在飞行员飞行训练中使用扩展现实(XR)频谱技术作为传统模拟器的替代品已经引起了人们的极大兴趣。我们进行了一项系统评价,以评估XR技术在这方面的功效,并更好地了解这种使用的激励因素。系统评价遵循QUOROM框架(适用于教育研究),筛选1237篇候选文章,其中67篇符合主题分析标准,其中5篇也符合元分析标准。现有文献强调了这些技术的好处,特别是由于沉浸式和空间意识,使更多现代教育理论的应用成为可能。虽然现有的文献是有关行业的大部分,有一个特别的重点是通用航空和更多的从头开始的飞行技能。meta分析结果表明,飞行员绩效有所改善,总体meta分析效应量估计为0.884 (z = 2.248, P = 0.025),具有统计学显著性,且中等强度。本综述的研究结果表明支持XR在飞行员飞行训练模拟器中的使用和意图。然而,存在许多严重的研究空白,例如模拟器病和晕动病的发生率可能更高,以及缺乏检查整个飞行员技能和课程背景下培训转移的强有力的研究试验。这项新颖的系统综述和荟萃分析代表了一项重要的尝试,旨在塑造和指导更好的研究,以帮助指导在飞行员短缺和航空业增长日益严重的情况下蓬勃发展的技术XR发展。
{"title":"Applications of extended reality in pilot flight simulator training: a systematic review with meta-analysis.","authors":"Alexander Somerville, Keith Joiner, Timothy Lynar, Graham Wild","doi":"10.1186/s42492-025-00206-w","DOIUrl":"10.1186/s42492-025-00206-w","url":null,"abstract":"<p><p>The use of extended reality (XR) spectrum technologies as substitutes to augment traditional simulators in pilot flight training has received significant interest in recent times. A systematic review was conducted to evaluate the efficacy of XR technologies for this purpose and better understand the motivating factors for this use. The systematic review followed the QUOROM framework (adapted for educational studies), screening 1237 candidate articles to 67 eligible for thematic analysis, with 5 of these also meeting meta-analysis criteria. Existing literature emphasizes the benefits of these technologies, particularly as a result of immersiveness and spatial awareness, enabling the application of more modern educational theories. Although the existing literature is concerned with much of the industry, there is a specific focus on general aviation and the more ab initio skills of flight. The results of the meta-analysis indicate improvements in pilot performance, with an overall meta-analytic effect size estimate of 0.884 (z = 2.248, P = 0.025), which is positive, statistically significant, and moderately strong. The findings of this review indicate support for the use and intention for the use of XR in pilot flight training simulators. However, multiple serious research gaps exist, such as the potential higher occurrence of simulator sickness and cybersickness, and a lack of robust research trials that examine transfer of training across the full pilot skill set and curricular contexts. This novel systematic review and meta-analysis represent a significant attempt to shape and direct better research to help to direct flourishing technological XR development in a time of increasing pilot shortages and aviation growth.</p>","PeriodicalId":29931,"journal":{"name":"Visual Computing for Industry Biomedicine and Art","volume":"8 1","pages":"25"},"PeriodicalIF":6.0,"publicationDate":"2025-10-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12546163/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145348911","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-02DOI: 10.1186/s42492-025-00205-x
Yiping Wu, Yue Li, Eugene Ch'ng, Jiaxin Gao, Tao Hong
Gesture-based interactions in a virtual reality (VR) setting can enhance our experience of traditional practices as part of preserving and communicating heritage. Cultural experiences embodied within VR environments are suggested to be an effective approach for experiencing intangible cultural heritage. Ceremonies, rituals, and related ancestral enactments are important for preserving cultural heritage. Kāi Bǐ Lǐ, also known as the First Writing Ceremony, is traditionally held for Chinese children before their first year of elementary school. However, gesture-based immersive VR for learning this tradition is new, and have not been developed within the community. This study focused on how users experienced learning cultural practices using gesture-based interactive VR across different age groups and hardware platforms. We first conducted an experiment with 60 participants (30 young adults and 30 children) using the First Writing Ceremony as a case study in which gestural interactions were elicited, designed, implemented, and evaluated. The study showed significant differences in play time and presence between the head-mounted display VR and desktop VR. In addition, children were less likely to experience fatigue than young adults. Following this, we conducted another study after eight months to investigate the VR systems' long-term learning effectiveness. This showed that children outperformed young adults in demonstrating greater knowledge retention. Our results and findings contribute to the design of gesture-based VR for different age groups across different platforms for experiencing, learning, and practicing cultural activities.
虚拟现实(VR)环境中基于手势的互动可以增强我们对传统习俗的体验,作为保护和传播遗产的一部分。在VR环境中体现文化体验是体验非物质文化遗产的有效途径。仪式、仪式和相关的祖传法令对保护文化遗产很重要。Kāi b / l /,也被称为初笔礼,传统上是为中国孩子在小学一年级之前举行的。然而,用于学习这一传统的基于手势的沉浸式VR是新的,并且尚未在社区中开发出来。这项研究的重点是用户如何在不同年龄组和硬件平台上使用基于手势的交互式VR体验学习文化实践。我们首先对60名参与者(30名年轻人和30名儿童)进行了实验,以“第一次书写仪式”为例研究手势互动的引发、设计、实施和评估。研究显示,头戴式虚拟现实和桌面虚拟现实在游戏时间和存在感上存在显著差异。此外,儿童比年轻人更不容易感到疲劳。在此之后,我们在8个月后进行了另一项研究,以调查VR系统的长期学习效果。这表明,儿童在知识记忆方面比年轻人表现得更好。我们的研究结果和发现有助于设计基于手势的VR,适用于不同年龄段、不同平台的体验、学习和实践文化活动。
{"title":"KaiBiLi: gesture-based immersive virtual reality ceremony for traditional Chinese cultural activities.","authors":"Yiping Wu, Yue Li, Eugene Ch'ng, Jiaxin Gao, Tao Hong","doi":"10.1186/s42492-025-00205-x","DOIUrl":"10.1186/s42492-025-00205-x","url":null,"abstract":"<p><p>Gesture-based interactions in a virtual reality (VR) setting can enhance our experience of traditional practices as part of preserving and communicating heritage. Cultural experiences embodied within VR environments are suggested to be an effective approach for experiencing intangible cultural heritage. Ceremonies, rituals, and related ancestral enactments are important for preserving cultural heritage. Kāi Bǐ Lǐ, also known as the First Writing Ceremony, is traditionally held for Chinese children before their first year of elementary school. However, gesture-based immersive VR for learning this tradition is new, and have not been developed within the community. This study focused on how users experienced learning cultural practices using gesture-based interactive VR across different age groups and hardware platforms. We first conducted an experiment with 60 participants (30 young adults and 30 children) using the First Writing Ceremony as a case study in which gestural interactions were elicited, designed, implemented, and evaluated. The study showed significant differences in play time and presence between the head-mounted display VR and desktop VR. In addition, children were less likely to experience fatigue than young adults. Following this, we conducted another study after eight months to investigate the VR systems' long-term learning effectiveness. This showed that children outperformed young adults in demonstrating greater knowledge retention. Our results and findings contribute to the design of gesture-based VR for different age groups across different platforms for experiencing, learning, and practicing cultural activities.</p>","PeriodicalId":29931,"journal":{"name":"Visual Computing for Industry Biomedicine and Art","volume":"8 1","pages":"24"},"PeriodicalIF":6.0,"publicationDate":"2025-10-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12491145/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145207799","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01DOI: 10.1186/s42492-025-00201-1
Loufei Guo, Shuaitong Zhang, Hongbo Chen, Yifu Li, Yang Liu, Wancheng Liu, Qiang Wang, Zhenchao Tang, Ping Jiang, Junjie Wang
In recent years, the application of artificial intelligence (AI) in medical image analysis has drawn increasing attention in clinical studies of gynecologic tumors. This study presents the development and prospects of AI applications to assist in the treatment of gynecological oncology. The Web of Science database was screened for articles published until August 2023. "artificial intelligence," "deep learning," "machine learning," "radiomics," "radiotherapy," "chemoradiotherapy," "neoadjuvant therapy," "immunotherapy," "gynecological malignancy," "cervical carcinoma," "cervical cancer," "ovarian cancer," "endometrial cancer," "vulvar cancer," "Vaginal cancer" were used as keywords. Research articles related to AI-assisted treatment of gynecological cancers were included. A total of 317 articles were retrieved based on the search strategy, and 133 were selected by applying the inclusion and exclusion criteria, including 114 on cervical cancer, 10 on endometrial cancer, and 9 on ovarian cancer. Among the included studies, 44 (33%) focused on prognosis prediction, 24 (18%) on treatment response prediction, 13 (10%) on adverse event prediction, five (4%) on dose distribution prediction, and 47 (35%) on target volume delineation. Target volume delineation and dose prediction were performed using deep Learning methods. For the prediction of treatment response, prognosis, and adverse events, 57 studies (70%) used conventional radiomics methods, 13 (16%) used deep Learning methods, 8 (10%) used spatial-related unconventional radiomics methods, and 3 (4%) used temporal-related unconventional radiomics methods. In cervical and endometrial cancers, target prediction mostly included treatment response, overall survival, recurrence, toxicity undergoing radiotherapy, lymph node metastasis, and dose distribution. For ovarian cancer, the target prediction included platinum sensitivity and postoperative complications. The majority of the studies were single-center, retrospective, and small-scale; 101 studies (76%) had single-center data, 125 studies (94%) were retrospective, and 127 studies (95%) included Less than 500 cases. The application of AI in assisting treatment in gynecological oncology remains limited. Although the results of AI in predicting the response, prognosis, adverse events, and dose distribution in gynecological oncology are superior, it is evident that there is no validation of substantial data from multiple centers for these tasks.
近年来,人工智能(AI)在医学图像分析中的应用在妇科肿瘤的临床研究中越来越受到重视。本研究介绍了人工智能在辅助妇科肿瘤治疗中的应用进展和前景。Web of Science数据库筛选了2023年8月之前发表的文章。关键词是“人工智能”、“深度学习”、“机器学习”、“放射组学”、“放疗”、“放化疗”、“新辅助治疗”、“免疫治疗”、“妇科恶性肿瘤”、“宫颈癌”、“宫颈癌”、“卵巢癌”、“子宫内膜癌”、“外阴癌”、“阴道癌”。纳入了与人工智能辅助妇科癌症治疗相关的研究文章。根据检索策略共检索到317篇文献,应用纳入和排除标准筛选出133篇文献,其中宫颈癌114篇、子宫内膜癌10篇、卵巢癌9篇。纳入的研究中,44项(33%)研究集中于预后预测,24项(18%)研究集中于治疗反应预测,13项(10%)研究集中于不良事件预测,5项(4%)研究集中于剂量分布预测,47项(35%)研究集中于靶体积描绘。使用深度学习方法进行靶体积描绘和剂量预测。对于治疗反应、预后和不良事件的预测,57项研究(70%)使用常规放射组学方法,13项(16%)使用深度学习方法,8项(10%)使用与空间相关的非常规放射组学方法,3项(4%)使用与时间相关的非常规放射组学方法。在宫颈癌和子宫内膜癌中,目标预测主要包括治疗反应、总生存期、复发、放疗毒性、淋巴结转移和剂量分布。对于卵巢癌,目标预测包括铂敏感性和术后并发症。大多数研究为单中心、回顾性和小规模研究;101项研究(76%)为单中心数据,125项研究(94%)为回顾性研究,127项研究(95%)纳入病例少于500例。人工智能在妇科肿瘤辅助治疗中的应用仍然有限。虽然人工智能在预测妇科肿瘤的反应、预后、不良事件和剂量分布方面的结果是优越的,但很明显,没有来自多个中心的大量数据验证这些任务。
{"title":"Application of artificial intelligence in assisting treatment of gynecologic tumors: a systematic review.","authors":"Loufei Guo, Shuaitong Zhang, Hongbo Chen, Yifu Li, Yang Liu, Wancheng Liu, Qiang Wang, Zhenchao Tang, Ping Jiang, Junjie Wang","doi":"10.1186/s42492-025-00201-1","DOIUrl":"10.1186/s42492-025-00201-1","url":null,"abstract":"<p><p>In recent years, the application of artificial intelligence (AI) in medical image analysis has drawn increasing attention in clinical studies of gynecologic tumors. This study presents the development and prospects of AI applications to assist in the treatment of gynecological oncology. The Web of Science database was screened for articles published until August 2023. \"artificial intelligence,\" \"deep learning,\" \"machine learning,\" \"radiomics,\" \"radiotherapy,\" \"chemoradiotherapy,\" \"neoadjuvant therapy,\" \"immunotherapy,\" \"gynecological malignancy,\" \"cervical carcinoma,\" \"cervical cancer,\" \"ovarian cancer,\" \"endometrial cancer,\" \"vulvar cancer,\" \"Vaginal cancer\" were used as keywords. Research articles related to AI-assisted treatment of gynecological cancers were included. A total of 317 articles were retrieved based on the search strategy, and 133 were selected by applying the inclusion and exclusion criteria, including 114 on cervical cancer, 10 on endometrial cancer, and 9 on ovarian cancer. Among the included studies, 44 (33%) focused on prognosis prediction, 24 (18%) on treatment response prediction, 13 (10%) on adverse event prediction, five (4%) on dose distribution prediction, and 47 (35%) on target volume delineation. Target volume delineation and dose prediction were performed using deep Learning methods. For the prediction of treatment response, prognosis, and adverse events, 57 studies (70%) used conventional radiomics methods, 13 (16%) used deep Learning methods, 8 (10%) used spatial-related unconventional radiomics methods, and 3 (4%) used temporal-related unconventional radiomics methods. In cervical and endometrial cancers, target prediction mostly included treatment response, overall survival, recurrence, toxicity undergoing radiotherapy, lymph node metastasis, and dose distribution. For ovarian cancer, the target prediction included platinum sensitivity and postoperative complications. The majority of the studies were single-center, retrospective, and small-scale; 101 studies (76%) had single-center data, 125 studies (94%) were retrospective, and 127 studies (95%) included Less than 500 cases. The application of AI in assisting treatment in gynecological oncology remains limited. Although the results of AI in predicting the response, prognosis, adverse events, and dose distribution in gynecological oncology are superior, it is evident that there is no validation of substantial data from multiple centers for these tasks.</p>","PeriodicalId":29931,"journal":{"name":"Visual Computing for Industry Biomedicine and Art","volume":"8 1","pages":"23"},"PeriodicalIF":6.0,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12484435/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145201619","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Postoperative venous thromboembolism (VTE) in colorectal cancer (CRC) patients undergoing surgery results in poor prognosis. However, there are no effective tools for early screening and predicting VTE. In this study, we developed a machine learning (ML)-based model for predicting the risk of VTE following CRC surgery and tested its performance using an external dataset. A total of 3227 CRC surgery patients were enrolled from the Affiliated Hospital of Qingdao University and Yantai Yuhuangding Hospital (from January 2016 to December 2023). Subsequently, 1596 patients from the Affiliated Hospital of Qingdao University were assigned to the training set, and 716 patients from Yantai Yuhuangding Hospital were assigned to the external validation set. A model was developed and trained using six ML algorithms using the stacking ensemble technique. Moreover, all models were developed using the tenfold cross-validation on the training set, and their performance was tested using an independent external validation set. In the training set, 173 (10.8%) patients developed VTE, 163 (10.2%) patients experienced deep venous thrombosis, and 29 (1.82%) cases had pulmonary embolism (PE). In the external validation set, 85 (11.9%) cases of VTE, 83 (11.6%) cases of deep vein thrombosis, and 14 (1.96%) cases of PE were recorded. The analysis revealed that the stacking model outperformed all other models in the external validation set, achieving significantly better performance in all metrics: the area under the receiver operating characteristic curve = 0.840 (0.790-0.887), accuracy = 0.810 (0.783-0.836), specificity = 0.819 (0.790-0.848), sensitivity = 0.741 (0.652-0.825), and recall = 0.959 (0.942-0.975). The stacking model for surgical CRC patients shows promise in enabling timely clinical detection of high-risk cases. This method facilitates the prioritized implementation of prophylactic anticoagulation in confirmed high-risk individuals, thereby mitigating unnecessary pharmacological intervention in low-risk populations.
{"title":"Development and validation of a machine learning model for predicting venous thromboembolism complications following colorectal cancer surgery.","authors":"Zongsheng Sun, Di Hao, Mingyu Yang, Wenzhi Wu, Hanhui Jing, Zhensong Yang, Anbang Sun, Wentao Xie, Longbo Zheng, Xixun Wang, Dongsheng Wang, Yun Lu, Guangye Tian, Shanglong Liu","doi":"10.1186/s42492-025-00204-y","DOIUrl":"10.1186/s42492-025-00204-y","url":null,"abstract":"<p><p>Postoperative venous thromboembolism (VTE) in colorectal cancer (CRC) patients undergoing surgery results in poor prognosis. However, there are no effective tools for early screening and predicting VTE. In this study, we developed a machine learning (ML)-based model for predicting the risk of VTE following CRC surgery and tested its performance using an external dataset. A total of 3227 CRC surgery patients were enrolled from the Affiliated Hospital of Qingdao University and Yantai Yuhuangding Hospital (from January 2016 to December 2023). Subsequently, 1596 patients from the Affiliated Hospital of Qingdao University were assigned to the training set, and 716 patients from Yantai Yuhuangding Hospital were assigned to the external validation set. A model was developed and trained using six ML algorithms using the stacking ensemble technique. Moreover, all models were developed using the tenfold cross-validation on the training set, and their performance was tested using an independent external validation set. In the training set, 173 (10.8%) patients developed VTE, 163 (10.2%) patients experienced deep venous thrombosis, and 29 (1.82%) cases had pulmonary embolism (PE). In the external validation set, 85 (11.9%) cases of VTE, 83 (11.6%) cases of deep vein thrombosis, and 14 (1.96%) cases of PE were recorded. The analysis revealed that the stacking model outperformed all other models in the external validation set, achieving significantly better performance in all metrics: the area under the receiver operating characteristic curve = 0.840 (0.790-0.887), accuracy = 0.810 (0.783-0.836), specificity = 0.819 (0.790-0.848), sensitivity = 0.741 (0.652-0.825), and recall = 0.959 (0.942-0.975). The stacking model for surgical CRC patients shows promise in enabling timely clinical detection of high-risk cases. This method facilitates the prioritized implementation of prophylactic anticoagulation in confirmed high-risk individuals, thereby mitigating unnecessary pharmacological intervention in low-risk populations.</p>","PeriodicalId":29931,"journal":{"name":"Visual Computing for Industry Biomedicine and Art","volume":"8 1","pages":"22"},"PeriodicalIF":6.0,"publicationDate":"2025-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12425853/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145041587","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-09-03DOI: 10.1186/s42492-025-00203-z
Graham Wild, Aziida Nanyonga, Anam Iqbal, Shehar Bano, Alexander Somerville, Luke Pollock
This review examines the current applications, benefits, challenges, and future potential of artificial intelligence (AI) and immersive aviation technologies. AI has been applied across various domains, including flight operations, air traffic control, maintenance, and ground handling. AI enhances aviation safety by enabling pilot assistance systems, mitigating human error, streamlining safety management systems, and aiding in accident analysis. Lightweight AI models are crucial for mobile applications in aviation, particularly for resource-constrained environments such as drones. Hardware considerations involve trade-offs between energy-efficient field-programmable gate arrays and power-consuming graphics processing units. Battery and thermal management are critical for mobile device applications. Although AI integration has numerous benefits, including enhanced safety, improved efficiency, and reduced environmental impact, it also presents challenges. Addressing algorithmic bias, ensuring cybersecurity, and managing the relationship between human operators and AI systems are crucial. The future of aviation will likely involve even more sophisticated AI algorithms, advanced hardware, and increased integration of AI with augmented reality and virtual reality, creating new possibilities for training and operations, and ultimately leading to a safer, more efficient, and more sustainable aviation industry.
{"title":"Lightweight and mobile artificial intelligence and immersive technologies in aviation.","authors":"Graham Wild, Aziida Nanyonga, Anam Iqbal, Shehar Bano, Alexander Somerville, Luke Pollock","doi":"10.1186/s42492-025-00203-z","DOIUrl":"10.1186/s42492-025-00203-z","url":null,"abstract":"<p><p>This review examines the current applications, benefits, challenges, and future potential of artificial intelligence (AI) and immersive aviation technologies. AI has been applied across various domains, including flight operations, air traffic control, maintenance, and ground handling. AI enhances aviation safety by enabling pilot assistance systems, mitigating human error, streamlining safety management systems, and aiding in accident analysis. Lightweight AI models are crucial for mobile applications in aviation, particularly for resource-constrained environments such as drones. Hardware considerations involve trade-offs between energy-efficient field-programmable gate arrays and power-consuming graphics processing units. Battery and thermal management are critical for mobile device applications. Although AI integration has numerous benefits, including enhanced safety, improved efficiency, and reduced environmental impact, it also presents challenges. Addressing algorithmic bias, ensuring cybersecurity, and managing the relationship between human operators and AI systems are crucial. The future of aviation will likely involve even more sophisticated AI algorithms, advanced hardware, and increased integration of AI with augmented reality and virtual reality, creating new possibilities for training and operations, and ultimately leading to a safer, more efficient, and more sustainable aviation industry.</p>","PeriodicalId":29931,"journal":{"name":"Visual Computing for Industry Biomedicine and Art","volume":"8 1","pages":"21"},"PeriodicalIF":6.0,"publicationDate":"2025-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12408884/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144972036","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-09-01DOI: 10.1186/s42492-025-00202-0
Chen Bai, Tan Li, Yanyan Zheng, Gang Yuan, Jian Zheng, Hui Zhao
Post-stroke cognitive impairment (PSCI) is a common and debilitating consequence of stroke that often arises from complex interactions between diverse brain alterations. The accurate early prediction of PSCI is critical for guiding personalized interventions. However, existing methods often struggle to capture complex structural disruptions and integrate multimodal information effectively. This study proposes the multimodal dynamic hierarchical clustering network (MDHCNet), a graph neural network designed for accurate and interpretable PSCI prediction. MDHCNet constructs brain graphs from diffusion-weighted imaging, magnetic resonance angiography, and T1- and T2-weighted images and integrates them with clinical features using a hierarchical cross-modal fusion module. Experimental results using a real-world stroke cohort demonstrated that MDHCNet consistently outperformed deep learning baselines. Ablation studies validated the benefits of multimodal fusion, while saliency-based interpretation highlighted discriminative brain regions associated with cognitive decline. These findings suggest that MDHCNet is an effective and explainable tool for early PSCI prediction, with the potential to support individualized clinical decision-making in stroke rehabilitation.
{"title":"Multimodal dynamic hierarchical clustering model for post-stroke cognitive impairment prediction.","authors":"Chen Bai, Tan Li, Yanyan Zheng, Gang Yuan, Jian Zheng, Hui Zhao","doi":"10.1186/s42492-025-00202-0","DOIUrl":"10.1186/s42492-025-00202-0","url":null,"abstract":"<p><p>Post-stroke cognitive impairment (PSCI) is a common and debilitating consequence of stroke that often arises from complex interactions between diverse brain alterations. The accurate early prediction of PSCI is critical for guiding personalized interventions. However, existing methods often struggle to capture complex structural disruptions and integrate multimodal information effectively. This study proposes the multimodal dynamic hierarchical clustering network (MDHCNet), a graph neural network designed for accurate and interpretable PSCI prediction. MDHCNet constructs brain graphs from diffusion-weighted imaging, magnetic resonance angiography, and T1- and T2-weighted images and integrates them with clinical features using a hierarchical cross-modal fusion module. Experimental results using a real-world stroke cohort demonstrated that MDHCNet consistently outperformed deep learning baselines. Ablation studies validated the benefits of multimodal fusion, while saliency-based interpretation highlighted discriminative brain regions associated with cognitive decline. These findings suggest that MDHCNet is an effective and explainable tool for early PSCI prediction, with the potential to support individualized clinical decision-making in stroke rehabilitation.</p>","PeriodicalId":29931,"journal":{"name":"Visual Computing for Industry Biomedicine and Art","volume":"8 1","pages":"20"},"PeriodicalIF":6.0,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12401840/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144972010","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-15DOI: 10.1186/s42492-025-00199-6
Xue Lu, Haoyan Zhang, Hidekatsu Kuroda, Matteo Garcovich, Victor de Ledinghen, Ivica Grgurević, Runze Linghu, Hong Ding, Jiandong Chang, Min Wu, Cheng Feng, Xinping Ren, Changzhu Liu, Tao Song, Fankun Meng, Yao Zhang, Ye Fang, Sumei Ma, Jinfen Wang, Xiaolong Qi, Jie Tian, Xin Yang, Jie Ren, Ping Liang, Kun Wang
Accurate, noninvasive diagnosis of compensated advanced chronic liver disease (cACLD) is essential for effective clinical management but remains challenging. This study aimed to develop a deep learning-based radiomics model using international multicenter data and to evaluate its performance by comparing it to the two-dimensional shear wave elastography (2D-SWE) cut-off method covering multiple countries or regions, etiologies, and ultrasound device manufacturers. This retrospective study included 1937 adult patients with chronic liver disease due to hepatitis B, hepatitis C, or metabolic dysfunction-associated steatotic liver disease. All patients underwent 2D-SWE imaging and liver biopsy at 17 centers across China, Japan, and Europe using devices from three manufacturers (SuperSonic Imagine, General Electric, and Mindray). The proposed generalized deep learning radiomics of elastography model integrated both elastographic images and liver stiffness measurements and was trained and tested on stratified internal and external datasets. A total of 1937 patients with 9472 2D-SWE images were included in the statistical analysis. Compared to 2D-SWE, the model achieved a higher area under the receiver operating characteristic curve (AUC) (0.89 vs 0.83, P = 0.025). It also achieved a highly consistent diagnosis across all subanalyses (P values: 0.21-0.91), whereas 2D-SWE exhibited different AUCs in the country or region (P < 0.001) and etiology (P = 0.005) subanalyses but not in the manufacturer subanalysis (P = 0.24). The model demonstrated more accurate and robust performance in noninvasive cACLD diagnosis than 2D-SWE across different countries or regions, etiologies, and manufacturers.
代偿性晚期慢性肝病(cACLD)的准确、无创诊断对于有效的临床治疗至关重要,但仍然具有挑战性。本研究旨在利用国际多中心数据开发基于深度学习的放射组学模型,并通过将其与涵盖多个国家或地区、病因和超声设备制造商的二维剪切波弹性成像(2D-SWE)截止方法进行比较来评估其性能。这项回顾性研究纳入了1937例由乙型肝炎、丙型肝炎或代谢功能障碍相关的脂肪变性肝病引起的慢性肝病成年患者。所有患者在中国、日本和欧洲的17个中心接受了2D-SWE成像和肝脏活检,使用的设备来自三家制造商(SuperSonic Imagine、General Electric和Mindray)。所提出的广义深度学习放射组学弹性成像模型集成了弹性成像图像和肝脏刚度测量,并在分层的内部和外部数据集上进行了训练和测试。统计分析共纳入1937例患者9472张2D-SWE图像。与2D-SWE相比,该模型在受者工作特征曲线(AUC)下的面积更高(0.89 vs 0.83, P = 0.025)。它还在所有亚分析中实现了高度一致的诊断(P值:0.21-0.91),而2D-SWE在国家或地区表现出不同的auc (P
{"title":"Deep learning radiomics of elastography for diagnosing compensated advanced chronic liver disease: an international multicenter study.","authors":"Xue Lu, Haoyan Zhang, Hidekatsu Kuroda, Matteo Garcovich, Victor de Ledinghen, Ivica Grgurević, Runze Linghu, Hong Ding, Jiandong Chang, Min Wu, Cheng Feng, Xinping Ren, Changzhu Liu, Tao Song, Fankun Meng, Yao Zhang, Ye Fang, Sumei Ma, Jinfen Wang, Xiaolong Qi, Jie Tian, Xin Yang, Jie Ren, Ping Liang, Kun Wang","doi":"10.1186/s42492-025-00199-6","DOIUrl":"10.1186/s42492-025-00199-6","url":null,"abstract":"<p><p>Accurate, noninvasive diagnosis of compensated advanced chronic liver disease (cACLD) is essential for effective clinical management but remains challenging. This study aimed to develop a deep learning-based radiomics model using international multicenter data and to evaluate its performance by comparing it to the two-dimensional shear wave elastography (2D-SWE) cut-off method covering multiple countries or regions, etiologies, and ultrasound device manufacturers. This retrospective study included 1937 adult patients with chronic liver disease due to hepatitis B, hepatitis C, or metabolic dysfunction-associated steatotic liver disease. All patients underwent 2D-SWE imaging and liver biopsy at 17 centers across China, Japan, and Europe using devices from three manufacturers (SuperSonic Imagine, General Electric, and Mindray). The proposed generalized deep learning radiomics of elastography model integrated both elastographic images and liver stiffness measurements and was trained and tested on stratified internal and external datasets. A total of 1937 patients with 9472 2D-SWE images were included in the statistical analysis. Compared to 2D-SWE, the model achieved a higher area under the receiver operating characteristic curve (AUC) (0.89 vs 0.83, P = 0.025). It also achieved a highly consistent diagnosis across all subanalyses (P values: 0.21-0.91), whereas 2D-SWE exhibited different AUCs in the country or region (P < 0.001) and etiology (P = 0.005) subanalyses but not in the manufacturer subanalysis (P = 0.24). The model demonstrated more accurate and robust performance in noninvasive cACLD diagnosis than 2D-SWE across different countries or regions, etiologies, and manufacturers.</p>","PeriodicalId":29931,"journal":{"name":"Visual Computing for Industry Biomedicine and Art","volume":"8 1","pages":"19"},"PeriodicalIF":6.0,"publicationDate":"2025-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12354435/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144856587","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-07-16DOI: 10.1186/s42492-025-00200-2
Karim Dabbabi, Tijeni Delleji
Unmanned aerial vehicle (UAV) tracking is a critical task in surveillance, security, and autonomous navigation applications. In this study, we propose graph neural network-tracker (GNN-tracker), a novel GNN-based UAV tracking framework that effectively integrates graph-based spatial-temporal modelling, Transformer-based feature extraction, and multi-sensor fusion to enhance tracking robustness and accuracy. Unlike traditional tracking approaches, GNN-tracker dynamically constructs a spatiotemporal graph representation, improving identity consistency and reducing tracking errors under OCC-heavy scenarios. Experimental evaluations on optical, thermal, and fused UAV datasets demonstrate the superiority of GNN-tracker (fused) over state-of-the-art methods. The proposed model achieves multiple object tracking accuracy (MOTA) scores of 91.4% (fused), 89.1% (optical), and 86.3% (thermal), surpassing TransT by 8.9% in MOTA and 7.7% in higher order tracking accuracy (HOTA). The HOTA scores of 82.3% (fused), 80.1% (optical), and 78.7% (thermal) validate its strong object association capabilities, while its frames per second of 58.9 (fused), 56.8 (optical), and 54.3 (thermal) ensures real-time performance. Additionally, ablation studies confirm the essential role of graph-based modelling and multi-sensor fusion, with performance drops of up to 8.9% in MOTA when these components are removed. Thus, GNN-tracker (fused) offers a highly accurate, robust, and efficient UAV tracking solution, effectively addressing real-world challenges across diverse environmental conditions and multiple sensor modalities.
{"title":"Graph neural network-tracker: a graph neural network-based multi-sensor fusion framework for robust unmanned aerial vehicle tracking.","authors":"Karim Dabbabi, Tijeni Delleji","doi":"10.1186/s42492-025-00200-2","DOIUrl":"10.1186/s42492-025-00200-2","url":null,"abstract":"<p><p>Unmanned aerial vehicle (UAV) tracking is a critical task in surveillance, security, and autonomous navigation applications. In this study, we propose graph neural network-tracker (GNN-tracker), a novel GNN-based UAV tracking framework that effectively integrates graph-based spatial-temporal modelling, Transformer-based feature extraction, and multi-sensor fusion to enhance tracking robustness and accuracy. Unlike traditional tracking approaches, GNN-tracker dynamically constructs a spatiotemporal graph representation, improving identity consistency and reducing tracking errors under OCC-heavy scenarios. Experimental evaluations on optical, thermal, and fused UAV datasets demonstrate the superiority of GNN-tracker (fused) over state-of-the-art methods. The proposed model achieves multiple object tracking accuracy (MOTA) scores of 91.4% (fused), 89.1% (optical), and 86.3% (thermal), surpassing TransT by 8.9% in MOTA and 7.7% in higher order tracking accuracy (HOTA). The HOTA scores of 82.3% (fused), 80.1% (optical), and 78.7% (thermal) validate its strong object association capabilities, while its frames per second of 58.9 (fused), 56.8 (optical), and 54.3 (thermal) ensures real-time performance. Additionally, ablation studies confirm the essential role of graph-based modelling and multi-sensor fusion, with performance drops of up to 8.9% in MOTA when these components are removed. Thus, GNN-tracker (fused) offers a highly accurate, robust, and efficient UAV tracking solution, effectively addressing real-world challenges across diverse environmental conditions and multiple sensor modalities.</p>","PeriodicalId":29931,"journal":{"name":"Visual Computing for Industry Biomedicine and Art","volume":"8 1","pages":"18"},"PeriodicalIF":3.2,"publicationDate":"2025-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12267811/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144643753","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-07-15DOI: 10.1186/s42492-025-00197-8
Asmaa Jittou, Khalid El Fazazy, Jamal Riffi
Placental segmentation is critical for the quantitative analysis of prenatal imaging applications. However, segmenting the placenta using magnetic resonance imaging (MRI) and ultrasound is challenging because of variations in fetal position, dynamic placental development, and image quality. Most segmentation methods define regions of interest with different shapes and intensities, encompassing the entire placenta or specific structures. Recently, deep learning has emerged as a key approach that offer high segmentation performance across diverse datasets. This review focuses on the recent advances in deep learning techniques for placental segmentation in medical imaging, specifically MRI and ultrasound modalities, and cover studies from 2019 to 2024. This review synthesizes recent research, expand knowledge in this innovative area, and highlight the potential of deep learning approaches to significantly enhance prenatal diagnostics. These findings emphasize the importance of selecting appropriate imaging modalities and model architectures tailored to specific clinical scenarios. In addition, integrating both MRI and ultrasound can enhance segmentation performance by leveraging complementary information. This review also discusses the challenges associated with the high costs and limited availability of advanced imaging technologies. It provides insights into the current state of placental segmentation techniques and their implications for improving maternal and fetal health outcomes, underscoring the transformative impact of deep learning on prenatal diagnostics.
{"title":"Placenta segmentation redefined: review of deep learning integration of magnetic resonance imaging and ultrasound imaging.","authors":"Asmaa Jittou, Khalid El Fazazy, Jamal Riffi","doi":"10.1186/s42492-025-00197-8","DOIUrl":"10.1186/s42492-025-00197-8","url":null,"abstract":"<p><p>Placental segmentation is critical for the quantitative analysis of prenatal imaging applications. However, segmenting the placenta using magnetic resonance imaging (MRI) and ultrasound is challenging because of variations in fetal position, dynamic placental development, and image quality. Most segmentation methods define regions of interest with different shapes and intensities, encompassing the entire placenta or specific structures. Recently, deep learning has emerged as a key approach that offer high segmentation performance across diverse datasets. This review focuses on the recent advances in deep learning techniques for placental segmentation in medical imaging, specifically MRI and ultrasound modalities, and cover studies from 2019 to 2024. This review synthesizes recent research, expand knowledge in this innovative area, and highlight the potential of deep learning approaches to significantly enhance prenatal diagnostics. These findings emphasize the importance of selecting appropriate imaging modalities and model architectures tailored to specific clinical scenarios. In addition, integrating both MRI and ultrasound can enhance segmentation performance by leveraging complementary information. This review also discusses the challenges associated with the high costs and limited availability of advanced imaging technologies. It provides insights into the current state of placental segmentation techniques and their implications for improving maternal and fetal health outcomes, underscoring the transformative impact of deep learning on prenatal diagnostics.</p>","PeriodicalId":29931,"journal":{"name":"Visual Computing for Industry Biomedicine and Art","volume":"8 1","pages":"17"},"PeriodicalIF":3.2,"publicationDate":"2025-07-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12263505/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144638295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In human-robot collaborative tasks, human trust in robots can reduce resistance to them, thereby increasing the success rate of task execution. However, most existing studies have focused on improving the success rate of human-robot collaboration (HRC) rather than on enhancing collaboration efficiency. To improve the overall collaboration efficiency while maintaining a high success rate, this study proposes an active interaction strategy generation for HRC based on trust. First, a trust-based optimal robot strategy generation method was proposed to generate the robot's optimal strategy in a HRC. This method employs a tree to model the HRC process under different robot strategies and calculates the optimal strategy based on the modeling results for the robot to execute. Second, the robot's performance was evaluated to calculate human's trust in a robot. A robot performance evaluation method based on a visual language model was also proposed. The evaluation results were input into the trust model to compute human's current trust. Finally, each time an object operation was completed, the robot's performance evaluation and optimal strategy generation methods worked together to automatically generate the optimal strategy of the robot for the next step until the entire collaborative task was completed. The experimental results demonstrates that this method significantly improve collaborative efficiency, achieving a high success rate in HRC.
{"title":"Active interaction strategy generation for human-robot collaboration based on trust.","authors":"Yujie Guo, Pengfei Yi, Xiaopeng Wei, Dongsheng Zhou","doi":"10.1186/s42492-025-00198-7","DOIUrl":"10.1186/s42492-025-00198-7","url":null,"abstract":"<p><p>In human-robot collaborative tasks, human trust in robots can reduce resistance to them, thereby increasing the success rate of task execution. However, most existing studies have focused on improving the success rate of human-robot collaboration (HRC) rather than on enhancing collaboration efficiency. To improve the overall collaboration efficiency while maintaining a high success rate, this study proposes an active interaction strategy generation for HRC based on trust. First, a trust-based optimal robot strategy generation method was proposed to generate the robot's optimal strategy in a HRC. This method employs a tree to model the HRC process under different robot strategies and calculates the optimal strategy based on the modeling results for the robot to execute. Second, the robot's performance was evaluated to calculate human's trust in a robot. A robot performance evaluation method based on a visual language model was also proposed. The evaluation results were input into the trust model to compute human's current trust. Finally, each time an object operation was completed, the robot's performance evaluation and optimal strategy generation methods worked together to automatically generate the optimal strategy of the robot for the next step until the entire collaborative task was completed. The experimental results demonstrates that this method significantly improve collaborative efficiency, achieving a high success rate in HRC.</p>","PeriodicalId":29931,"journal":{"name":"Visual Computing for Industry Biomedicine and Art","volume":"8 1","pages":"16"},"PeriodicalIF":3.2,"publicationDate":"2025-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12185789/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144477081","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}