首页 > 最新文献

Visual Computing for Industry Biomedicine and Art最新文献

英文 中文
MetaChest: generalized few-shot learning of pathologies from chest X-rays. MetaChest:从胸部x光片中泛化的少量病理学习。
IF 6 4区 计算机科学 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-02-06 DOI: 10.1186/s42492-026-00214-4
Berenice Montalvo-Lezama, Gibran Fuentes-Pineda

The limited availability of annotated data presents a major challenge in applying deep learning methods to medical image analysis. Few-shot learning methods aim to recognize new classes from only a few labeled examples. These methods are typically investigated within a standard few-shot learning paradigm, in which all classes in a task are new. However, medical applications, such as pathology classification from chest X-rays, often require learning new classes while simultaneously leveraging the knowledge of previously known ones, a scenario more closely aligned with generalized few-shot classification. Despite its practical relevance, few-shot learning has rarely been investigated in this context. This study presents MetaChest, a large-scale dataset of 479,215 chest X-rays collected from four public databases. It includes a meta-set partition specifically designed for standard few-shot classification, as well as an algorithm for generating multi-label episodes. Extensive experiments were conducted to evaluate both the standard transfer learning (TL) approach and an extension of ProtoNet across a wide range of few-shot multi-label classification tasks. The results indicate that increasing the number of classes per episode and the number of training examples per class improves the classification performance. Notably, the TL approach consistently outperformed the ProtoNet extension, even though it was not specifically tailored for few-shot learning. Furthermore, higher-resolution images improved the accuracy at the cost of additional computation, whereas efficient model architectures achieved performances comparable to larger models with significantly reduced resource requirements.

注释数据的有限可用性是将深度学习方法应用于医学图像分析的主要挑战。few -shot学习方法旨在仅从少数标记的示例中识别新类。这些方法通常在标准的少量学习范式中进行研究,在这种范式中,任务中的所有类都是新的。然而,医学应用,如胸部x光的病理分类,通常需要学习新的课程,同时利用以前已知的知识,这种情况更接近于广义的几次分类。尽管它具有实际意义,但很少在此背景下进行研究。这项研究提出了MetaChest,这是一个从四个公共数据库收集的479,215张胸部x光片的大型数据集。它包括一个专门为标准少镜头分类设计的元集分区,以及一个用于生成多标签剧集的算法。我们进行了大量的实验,以评估标准迁移学习(TL)方法和ProtoNet的扩展在广泛的少量多标签分类任务中的应用。结果表明,增加每集的类数和每个类的训练样例数可以提高分类性能。值得注意的是,TL方法始终优于ProtoNet扩展,即使它不是专门为少量学习量身定制的。此外,更高分辨率的图像以额外的计算为代价提高了精度,而高效的模型架构在显著减少资源需求的情况下实现了与大型模型相当的性能。
{"title":"MetaChest: generalized few-shot learning of pathologies from chest X-rays.","authors":"Berenice Montalvo-Lezama, Gibran Fuentes-Pineda","doi":"10.1186/s42492-026-00214-4","DOIUrl":"10.1186/s42492-026-00214-4","url":null,"abstract":"<p><p>The limited availability of annotated data presents a major challenge in applying deep learning methods to medical image analysis. Few-shot learning methods aim to recognize new classes from only a few labeled examples. These methods are typically investigated within a standard few-shot learning paradigm, in which all classes in a task are new. However, medical applications, such as pathology classification from chest X-rays, often require learning new classes while simultaneously leveraging the knowledge of previously known ones, a scenario more closely aligned with generalized few-shot classification. Despite its practical relevance, few-shot learning has rarely been investigated in this context. This study presents MetaChest, a large-scale dataset of 479,215 chest X-rays collected from four public databases. It includes a meta-set partition specifically designed for standard few-shot classification, as well as an algorithm for generating multi-label episodes. Extensive experiments were conducted to evaluate both the standard transfer learning (TL) approach and an extension of ProtoNet across a wide range of few-shot multi-label classification tasks. The results indicate that increasing the number of classes per episode and the number of training examples per class improves the classification performance. Notably, the TL approach consistently outperformed the ProtoNet extension, even though it was not specifically tailored for few-shot learning. Furthermore, higher-resolution images improved the accuracy at the cost of additional computation, whereas efficient model architectures achieved performances comparable to larger models with significantly reduced resource requirements.</p>","PeriodicalId":29931,"journal":{"name":"Visual Computing for Industry Biomedicine and Art","volume":"9 1","pages":"4"},"PeriodicalIF":6.0,"publicationDate":"2026-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12876522/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146126940","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Advances in photoacoustic imaging reconstruction and quantitative analysis for biomedical applications. 生物医学领域光声成像重建与定量分析研究进展。
IF 6 4区 计算机科学 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-02-01 DOI: 10.1186/s42492-025-00213-x
Lei Wang, Weiming Zeng, Kai Long, Hongyu Chen, Rongfeng Lan, Li Liu, Wai Ting Siok, Nizhuan Wang

Photoacoustic imaging (PAI), a modality that combines the high contrast of optical imaging with the deep penetration of ultrasound, is rapidly transitioning from preclinical research to clinical practice. However, its widespread clinical adoption faces challenges such as the inherent trade-off between penetration depth and spatial resolution, along with the demand for faster imaging speeds. This review comprehensively examines the fundamental principles of PAI, focusing on three primary implementations: photoacoustic computed tomography, photoacoustic microscopy, and photoacoustic endoscopy. It critically analyzes their respective advantages and limitations to provide insights into practical applications. The discussion then extends to recent advancements in image reconstruction and artifact suppression, where both conventional and deep learning (DL)-based approaches have been highlighted for their role in enhancing image quality and streamlining workflows. Furthermore, this work explores progress in quantitative PAI, particularly its ability to precisely measure hemoglobin concentration, oxygen saturation, and other physiological biomarkers. Finally, this review outlines emerging trends and future directions, underscoring the transformative potential of DL in shaping the clinical evolution of PAI.

光声成像(PAI)是一种将光学成像的高对比度与超声的深穿透性相结合的技术,正迅速从临床前研究向临床实践过渡。然而,其广泛的临床应用面临着挑战,例如穿透深度和空间分辨率之间的内在权衡,以及对更快成像速度的需求。本文综述了PAI的基本原理,重点介绍了三种主要的实现方法:光声计算机断层扫描、光声显微镜和光声内窥镜检查。它批判性地分析了各自的优势和局限性,以提供对实际应用的见解。然后讨论扩展到图像重建和伪影抑制的最新进展,其中传统和基于深度学习(DL)的方法都因其在提高图像质量和简化工作流程方面的作用而得到强调。此外,本工作探讨了定量PAI的进展,特别是其精确测量血红蛋白浓度、氧饱和度和其他生理生物标志物的能力。最后,本综述概述了新兴趋势和未来方向,强调DL在塑造PAI临床演变中的变革潜力。
{"title":"Advances in photoacoustic imaging reconstruction and quantitative analysis for biomedical applications.","authors":"Lei Wang, Weiming Zeng, Kai Long, Hongyu Chen, Rongfeng Lan, Li Liu, Wai Ting Siok, Nizhuan Wang","doi":"10.1186/s42492-025-00213-x","DOIUrl":"10.1186/s42492-025-00213-x","url":null,"abstract":"<p><p>Photoacoustic imaging (PAI), a modality that combines the high contrast of optical imaging with the deep penetration of ultrasound, is rapidly transitioning from preclinical research to clinical practice. However, its widespread clinical adoption faces challenges such as the inherent trade-off between penetration depth and spatial resolution, along with the demand for faster imaging speeds. This review comprehensively examines the fundamental principles of PAI, focusing on three primary implementations: photoacoustic computed tomography, photoacoustic microscopy, and photoacoustic endoscopy. It critically analyzes their respective advantages and limitations to provide insights into practical applications. The discussion then extends to recent advancements in image reconstruction and artifact suppression, where both conventional and deep learning (DL)-based approaches have been highlighted for their role in enhancing image quality and streamlining workflows. Furthermore, this work explores progress in quantitative PAI, particularly its ability to precisely measure hemoglobin concentration, oxygen saturation, and other physiological biomarkers. Finally, this review outlines emerging trends and future directions, underscoring the transformative potential of DL in shaping the clinical evolution of PAI.</p>","PeriodicalId":29931,"journal":{"name":"Visual Computing for Industry Biomedicine and Art","volume":"9 1","pages":"3"},"PeriodicalIF":6.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12860771/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146097413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Development of an optically emulated computed tomography scanner for college education. 大学教育用光学模拟计算机断层扫描仪的研制。
IF 6 4区 计算机科学 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-01-16 DOI: 10.1186/s42492-025-00211-z
Md Motaleb Hossen Manik, William Muldowney, Md Zabirul Islam, Ge Wang

Computed tomography (CT) is a powerful imaging modality widely used in medicine, research, and industry for noninvasive visualization of internal structures. However, conventional CT systems rely on X-rays, which involve radiation exposure, high equipment costs, and complex regulatory requirements, making them unsuitable for educational or low-resource settings. To address these limitations, we developed a compact, low-cost, optically emulated CT scanner that uses visible light to image semi-transparent specimens. The system consists of a rotating stage enclosed within a light-isolated box, backlight illumination, and a fixed digital single-lens reflex camera. A Teensy 2.0 microcontroller regulates the rotation of the stage, while MATLAB is used to process the captured images using the inverse Radon transform and visualize the reconstructed volume using the Volumetric 3D MATLAB toolbox. Experimental results using a lemon slice demonstrate that the scanner can resolve internal features such as the peel, pulp, and seeds in both 2D and 3D renderings. This system offers a safe and affordable platform for demonstrating CT principles, with potential applications in education, industrial inspection, and visual computing.

计算机断层扫描(CT)是一种强大的成像方式,广泛应用于医学、研究和工业中,用于内部结构的无创可视化。然而,传统的CT系统依赖于x射线,这涉及到辐射暴露、高设备成本和复杂的监管要求,使其不适合教育或低资源环境。为了解决这些限制,我们开发了一种紧凑、低成本、光学模拟的CT扫描仪,它使用可见光对半透明标本进行成像。该系统由一个封闭在光隔离盒内的旋转舞台、背光照明和一个固定的数字单镜头反光相机组成。一个Teensy 2.0微控制器调节舞台的旋转,而MATLAB使用反Radon变换对捕获的图像进行处理,并使用Volumetric 3D MATLAB工具箱将重建的体积可视化。使用柠檬切片的实验结果表明,扫描仪可以在2D和3D渲染中解析内部特征,如果皮,果肉和种子。该系统为演示CT原理提供了一个安全且经济实惠的平台,在教育、工业检测和视觉计算方面具有潜在的应用前景。
{"title":"Development of an optically emulated computed tomography scanner for college education.","authors":"Md Motaleb Hossen Manik, William Muldowney, Md Zabirul Islam, Ge Wang","doi":"10.1186/s42492-025-00211-z","DOIUrl":"10.1186/s42492-025-00211-z","url":null,"abstract":"<p><p>Computed tomography (CT) is a powerful imaging modality widely used in medicine, research, and industry for noninvasive visualization of internal structures. However, conventional CT systems rely on X-rays, which involve radiation exposure, high equipment costs, and complex regulatory requirements, making them unsuitable for educational or low-resource settings. To address these limitations, we developed a compact, low-cost, optically emulated CT scanner that uses visible light to image semi-transparent specimens. The system consists of a rotating stage enclosed within a light-isolated box, backlight illumination, and a fixed digital single-lens reflex camera. A Teensy 2.0 microcontroller regulates the rotation of the stage, while MATLAB is used to process the captured images using the inverse Radon transform and visualize the reconstructed volume using the Volumetric 3D MATLAB toolbox. Experimental results using a lemon slice demonstrate that the scanner can resolve internal features such as the peel, pulp, and seeds in both 2D and 3D renderings. This system offers a safe and affordable platform for demonstrating CT principles, with potential applications in education, industrial inspection, and visual computing.</p>","PeriodicalId":29931,"journal":{"name":"Visual Computing for Industry Biomedicine and Art","volume":"9 1","pages":"2"},"PeriodicalIF":6.0,"publicationDate":"2026-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12808011/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145991049","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Artificial intelligence-aided assignment of journal submissions to associate editors-a feasibility study on IEEE transactions on medical imaging. 人工智能辅助期刊提交给副编辑的分配——IEEE医学成像交易的可行性研究。
IF 6 4区 计算机科学 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-01-12 DOI: 10.1186/s42492-025-00212-y
Xuanang Xu, Joshua Yan, Gloria Nwachukwu, Hongming Shan, Uwe Kruger, Ge Wang

Efficient and accurate assignment of journal submissions to suitable associate editors (AEs) is critical in maintaining review quality and timeliness, particularly in high-volume, rapidly evolving fields such as medical imaging. This study investigates the feasibility of leveraging large language models for AE-paper matching in IEEE Transactions on Medical Imaging. An AE database was curated from historical AE assignments and AE-authored publications, and extracted six key textual components from each paper title, four categories of structured keywords, and abstracts. ModernBERT was employed locally to generate high-dimensional semantic embeddings, which were then reduced using principal component analysis (PCA) for efficient similarity computation. Keyword similarity, derived from structured domain-specific metadata, and textual similarity from ModernBERT embeddings were combined to rank the candidate AEs. Experiments on internal (historical assignments) and external (AE Publications) test sets showed that keyword similarity is the dominant contributor to matching performance. Contrarily, textual similarity offers complementary gains, particularly when PCA is applied. Ablation studies confirmed that structured keywords alone provide strong matching accuracy, with titles offering additional benefits and abstracts offering minimal improvements. The proposed approach offers a practical, interpretable, and scalable tool for editorial workflows, reduces manual workload, and supports high-quality peer reviews.

有效和准确地将期刊投稿分配给合适的副编辑(ae)对于保持评审质量和及时性至关重要,特别是在医学成像等高容量、快速发展的领域。本研究探讨利用大型语言模型进行IEEE医学影像汇刊电子论文匹配的可行性。从AE的历史作业和AE撰写的出版物中整理出AE数据库,并从每篇论文标题、四类结构化关键词和摘要中提取出六个关键文本成分。利用ModernBERT局部生成高维语义嵌入,然后利用主成分分析(PCA)对其进行约简,实现高效的相似度计算。从结构化领域特定元数据中获得的关键字相似度和来自ModernBERT嵌入的文本相似度被结合起来对候选ae进行排名。在内部(历史作业)和外部(AE出版物)测试集上的实验表明,关键词相似度是影响匹配性能的主要因素。相反,文本相似度提供互补增益,特别是在应用PCA时。消融研究证实,结构化关键词本身提供了很强的匹配准确性,标题提供了额外的好处,摘要提供了最小的改进。所建议的方法为编辑工作流程提供了一个实用的、可解释的和可扩展的工具,减少了手工工作量,并支持高质量的同行评审。
{"title":"Artificial intelligence-aided assignment of journal submissions to associate editors-a feasibility study on IEEE transactions on medical imaging.","authors":"Xuanang Xu, Joshua Yan, Gloria Nwachukwu, Hongming Shan, Uwe Kruger, Ge Wang","doi":"10.1186/s42492-025-00212-y","DOIUrl":"10.1186/s42492-025-00212-y","url":null,"abstract":"<p><p>Efficient and accurate assignment of journal submissions to suitable associate editors (AEs) is critical in maintaining review quality and timeliness, particularly in high-volume, rapidly evolving fields such as medical imaging. This study investigates the feasibility of leveraging large language models for AE-paper matching in IEEE Transactions on Medical Imaging. An AE database was curated from historical AE assignments and AE-authored publications, and extracted six key textual components from each paper title, four categories of structured keywords, and abstracts. ModernBERT was employed locally to generate high-dimensional semantic embeddings, which were then reduced using principal component analysis (PCA) for efficient similarity computation. Keyword similarity, derived from structured domain-specific metadata, and textual similarity from ModernBERT embeddings were combined to rank the candidate AEs. Experiments on internal (historical assignments) and external (AE Publications) test sets showed that keyword similarity is the dominant contributor to matching performance. Contrarily, textual similarity offers complementary gains, particularly when PCA is applied. Ablation studies confirmed that structured keywords alone provide strong matching accuracy, with titles offering additional benefits and abstracts offering minimal improvements. The proposed approach offers a practical, interpretable, and scalable tool for editorial workflows, reduces manual workload, and supports high-quality peer reviews.</p>","PeriodicalId":29931,"journal":{"name":"Visual Computing for Industry Biomedicine and Art","volume":"9 1","pages":"1"},"PeriodicalIF":6.0,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12791093/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145953273","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Text-to-3D scene generation framework: bridging textual descriptions to high-fidelity 3D scenes. 文本到3D场景生成框架:桥接文本描述到高保真3D场景。
IF 6 4区 计算机科学 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-12-18 DOI: 10.1186/s42492-025-00210-0
Zuan Gu, Tianhan Gao, Huimin Liu

Text-to-3D scene generation is pivotal for digital content creation; however, existing methods often struggle with global consistency across views. We present 3DS-Gen, a modular "generate-then-reconstruct" framework that first produces a temporally coherent multi-view video prior and then reconstructs consistent 3D scenes using sparse geometry estimation and Gaussian optimization. A cascaded variational autoencoder (2D for spatial compression and 3D for temporal compression) provides a compact and coherent latent sequence that facilitates robust reconstruction. An adaptive density threshold improves detailed allocation in the Gaussian stage under a fixed computational budget. While explicit meshes can be extracted from the optimized representation when needed, our claims emphasize multiview consistency and reconstructability; the mesh quality depends on the video prior and the chosen explicitification backend. 3DS-Gen runs on a single GPU and yields coherent scene reconstructions across diverse prompts, thereby providing a practical bridge between text and 3D content creation.

文本到3d场景生成是数字内容创作的关键;然而,现有的方法经常与跨视图的全局一致性作斗争。我们提出了3DS-Gen,一个模块化的“生成-然后重建”框架,首先产生一个暂时连贯的多视图视频,然后使用稀疏几何估计和高斯优化重建一致的3D场景。级联变分自编码器(2D用于空间压缩,3D用于时间压缩)提供紧凑一致的潜在序列,促进鲁棒重建。在固定的计算预算下,自适应密度阈值改善了高斯阶段的详细分配。虽然可以在需要时从优化的表示中提取显式网格,但我们的要求强调多视图一致性和可重构性;网格质量取决于视频先验和选择的显式后端。3DS-Gen在单个GPU上运行,并在不同的提示中产生连贯的场景重建,从而在文本和3D内容创建之间提供实用的桥梁。
{"title":"Text-to-3D scene generation framework: bridging textual descriptions to high-fidelity 3D scenes.","authors":"Zuan Gu, Tianhan Gao, Huimin Liu","doi":"10.1186/s42492-025-00210-0","DOIUrl":"10.1186/s42492-025-00210-0","url":null,"abstract":"<p><p>Text-to-3D scene generation is pivotal for digital content creation; however, existing methods often struggle with global consistency across views. We present 3DS-Gen, a modular \"generate-then-reconstruct\" framework that first produces a temporally coherent multi-view video prior and then reconstructs consistent 3D scenes using sparse geometry estimation and Gaussian optimization. A cascaded variational autoencoder (2D for spatial compression and 3D for temporal compression) provides a compact and coherent latent sequence that facilitates robust reconstruction. An adaptive density threshold improves detailed allocation in the Gaussian stage under a fixed computational budget. While explicit meshes can be extracted from the optimized representation when needed, our claims emphasize multiview consistency and reconstructability; the mesh quality depends on the video prior and the chosen explicitification backend. 3DS-Gen runs on a single GPU and yields coherent scene reconstructions across diverse prompts, thereby providing a practical bridge between text and 3D content creation.</p>","PeriodicalId":29931,"journal":{"name":"Visual Computing for Industry Biomedicine and Art","volume":"8 1","pages":"29"},"PeriodicalIF":6.0,"publicationDate":"2025-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12712286/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145775790","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhanced temporal encoding-decoding for survival analysis of multimodal clinical data in smart healthcare. 智能医疗中多模式临床数据生存分析的增强时间编码解码。
IF 6 4区 计算机科学 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-12-12 DOI: 10.1186/s42492-025-00209-7
Xiaofeng Zhang, Zijie Pan, Yuhang Tian, Lili Wang, Tingting Xu, Li Chen, Xiangyun Liao, Tianyu Jiang

Effective survival analysis is essential for identifying optimal preventive treatments within smart healthcare systems and leveraging digital health advancements; however, existing prediction models face limitations, primarily relying on ensemble classification techniques with suboptimal performance in both target detection and predictive accuracy. To address these gaps, this paper proposes a multimodal framework that integrates enhanced facial feature detection and temporal predictive modeling. For facial feature extraction, this study developed a lightweight face-region convolutional neural network (FRegNet) specialized in detecting key facial components, such as eyes and lips in clinical patients that incorporates a residual backbone (Rstem) to enhance feature representation and a facial path aggregated feature pyramid network for multi-resolution feature fusion; comparative experiments reveal that FRegNet outperforms state-of-the-art target detection algorithms, achieving average precision (AP) of 0.922, average recall of 0.933, mean average precision (mAP) of 0.987, and precision of 0.98-significantly surpassing other mask region-based convolutional neural networks (RCNN) variants, such as mask RCNN-ResNeXt with AP of 0.789 and mAP of 0.957. Based on the extracted facial features and clinical physiological indicators, this study proposes an enhanced temporal encoding-decoding (ETED) model that integrates an adaptive attention mechanism and a gated weighting mechanism to improve predictive performance, with comparative results demonstrating that the ETED variant incorporating facial features (ETEncoding-Decoding-Face) outperforms traditional models, achieving an accuracy of 0.916, precision of 0.850, recall of 0.895, F1 of 0.884, and area under the curve (AUC) of 0.947-outperforming gradient boosting with an accuracy of 0.922, but AUC of 0.669, and other classifiers in comprehensive metrics. The results confirm that the multimodal dataset (facial features + physiological indicators) significantly enhances the prediction accuracy of the seven-day survival conditions of patients. Correlation analysis reveals that chronic health evaluation and mean arterial pressure are positively correlated with survival, while temperature, Glasgow Coma Scale, and fibrinogen are negatively correlated.

有效的生存分析对于在智能医疗系统中确定最佳预防治疗和利用数字健康进步至关重要;然而,现有的预测模型存在局限性,主要依赖于集成分类技术,在目标检测和预测精度方面都表现不佳。为了解决这些差距,本文提出了一个多模态框架,该框架集成了增强的面部特征检测和时间预测建模。在面部特征提取方面,本研究开发了一种轻量级的面部区域卷积神经网络(FRegNet),专门用于检测临床患者的关键面部成分,如眼睛和嘴唇,该网络结合了残馀骨干(system)来增强特征表征和面部路径聚合特征金字塔网络来进行多分辨率特征融合;对比实验表明,FRegNet优于最先进的目标检测算法,平均精度(AP)为0.922,平均查全率(recall)为0.933,平均平均精度(mAP)为0.987,精度为0.98,显著优于其他基于掩模区域的卷积神经网络(RCNN)变体,如掩模RCNN- resnext, AP为0.789,mAP为0.957。基于提取的面部特征和临床生理指标,本研究提出了一种集成自适应注意机制和门控加权机制的增强时间编码解码(etted)模型,以提高预测性能。对比结果表明,结合面部特征的etted变体(etenencoding -decoding - face)优于传统模型,准确率为0.916,精密度为0.850,召回率为0.895,F1为0.884。曲线下面积(AUC)为0.947,在综合指标上优于梯度提升,准确率为0.922,但AUC为0.669。结果证实,多模态数据集(面部特征+生理指标)显著提高了患者7天生存条件的预测精度。相关分析显示,慢性健康评价和平均动脉压与生存呈正相关,而体温、格拉斯哥昏迷量表和纤维蛋白原与生存负相关。
{"title":"Enhanced temporal encoding-decoding for survival analysis of multimodal clinical data in smart healthcare.","authors":"Xiaofeng Zhang, Zijie Pan, Yuhang Tian, Lili Wang, Tingting Xu, Li Chen, Xiangyun Liao, Tianyu Jiang","doi":"10.1186/s42492-025-00209-7","DOIUrl":"10.1186/s42492-025-00209-7","url":null,"abstract":"<p><p>Effective survival analysis is essential for identifying optimal preventive treatments within smart healthcare systems and leveraging digital health advancements; however, existing prediction models face limitations, primarily relying on ensemble classification techniques with suboptimal performance in both target detection and predictive accuracy. To address these gaps, this paper proposes a multimodal framework that integrates enhanced facial feature detection and temporal predictive modeling. For facial feature extraction, this study developed a lightweight face-region convolutional neural network (FRegNet) specialized in detecting key facial components, such as eyes and lips in clinical patients that incorporates a residual backbone (Rstem) to enhance feature representation and a facial path aggregated feature pyramid network for multi-resolution feature fusion; comparative experiments reveal that FRegNet outperforms state-of-the-art target detection algorithms, achieving average precision (AP) of 0.922, average recall of 0.933, mean average precision (mAP) of 0.987, and precision of 0.98-significantly surpassing other mask region-based convolutional neural networks (RCNN) variants, such as mask RCNN-ResNeXt with AP of 0.789 and mAP of 0.957. Based on the extracted facial features and clinical physiological indicators, this study proposes an enhanced temporal encoding-decoding (ETED) model that integrates an adaptive attention mechanism and a gated weighting mechanism to improve predictive performance, with comparative results demonstrating that the ETED variant incorporating facial features (ETEncoding-Decoding-Face) outperforms traditional models, achieving an accuracy of 0.916, precision of 0.850, recall of 0.895, F1 of 0.884, and area under the curve (AUC) of 0.947-outperforming gradient boosting with an accuracy of 0.922, but AUC of 0.669, and other classifiers in comprehensive metrics. The results confirm that the multimodal dataset (facial features + physiological indicators) significantly enhances the prediction accuracy of the seven-day survival conditions of patients. Correlation analysis reveals that chronic health evaluation and mean arterial pressure are positively correlated with survival, while temperature, Glasgow Coma Scale, and fibrinogen are negatively correlated.</p>","PeriodicalId":29931,"journal":{"name":"Visual Computing for Industry Biomedicine and Art","volume":"8 1","pages":"28"},"PeriodicalIF":6.0,"publicationDate":"2025-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12701133/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145744924","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comprehensive review of machine learning and deep learning techniques for epileptic seizure detection and prediction based on neuroimaging modalities. 基于神经成像模式的癫痫发作检测和预测的机器学习和深度学习技术综述。
IF 6 4区 计算机科学 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-12-11 DOI: 10.1186/s42492-025-00208-8
Khadija Slama, Ali Yahyaouy, Jamal Riffi, Mohamed Adnane Mahraz, Hamid Tairi

Epilepsy is a chronic neurological disorder characterized by recurrent seizures that can lead to death. Seizure treatment usually involves antiepileptic drugs and sometimes surgery, but patients with drug-resistant epilepsy often remain effectively untreated owing to the lack of targeted therapies. The development of a reliable technique for detecting and predicting epileptic seizures could significantly impact clinical treatment protocols and the care of patients with epilepsy. Over the years, researchers have developed various computational techniques using scalp electroencephalography (EEG), intracranial EEG, and other neuroimaging modalities, evolving from traditional signal processing methods (e.g., wavelet transforms and template matching) to advanced machine learning (ML, e.g., support vector machines and random forests) and deep learning (DL) algorithms (e.g., convolutional neural networks, recurrent neural networks, transformers, graph neural networks, and hybrid architectures). This review provides a detailed examination of epileptic seizure detection and prediction, covering the key aspects of signal processing, ML algorithms, and DL techniques applied to brainwave signals. We systematically categorized the techniques, analyzed key research trends, and identified critical challenges (e.g., data scarcity, model generalizability, and real-time processing). By highlighting the gaps in the literature, this review serves as a valuable resource for researchers and offers insights into future directions for improving the accuracy, interpretability, and clinical applicability of EEG-based seizure detection systems.

癫痫是一种慢性神经系统疾病,其特点是反复发作,可导致死亡。癫痫发作的治疗通常包括抗癫痫药物,有时还包括手术,但由于缺乏靶向治疗,耐药癫痫患者往往得不到有效治疗。开发一种可靠的检测和预测癫痫发作的技术可以显著影响临床治疗方案和癫痫患者的护理。多年来,研究人员利用头皮脑电图(EEG)、颅内脑电图(EEG)和其他神经成像模式开发了各种计算技术,从传统的信号处理方法(如小波变换和模板匹配)发展到先进的机器学习(ML,如支持向量机和随机森林)和深度学习(DL)算法(如卷积神经网络、循环神经网络、变压器、图神经网络、以及混合架构)。这篇综述提供了癫痫发作检测和预测的详细研究,涵盖了信号处理、ML算法和应用于脑波信号的DL技术的关键方面。我们系统地对这些技术进行了分类,分析了关键的研究趋势,并确定了关键的挑战(例如,数据稀缺性、模型通用性和实时处理)。通过强调文献中的空白,本综述为研究人员提供了宝贵的资源,并为提高基于脑电图的癫痫检测系统的准确性、可解释性和临床适用性提供了未来的方向。
{"title":"Comprehensive review of machine learning and deep learning techniques for epileptic seizure detection and prediction based on neuroimaging modalities.","authors":"Khadija Slama, Ali Yahyaouy, Jamal Riffi, Mohamed Adnane Mahraz, Hamid Tairi","doi":"10.1186/s42492-025-00208-8","DOIUrl":"10.1186/s42492-025-00208-8","url":null,"abstract":"<p><p>Epilepsy is a chronic neurological disorder characterized by recurrent seizures that can lead to death. Seizure treatment usually involves antiepileptic drugs and sometimes surgery, but patients with drug-resistant epilepsy often remain effectively untreated owing to the lack of targeted therapies. The development of a reliable technique for detecting and predicting epileptic seizures could significantly impact clinical treatment protocols and the care of patients with epilepsy. Over the years, researchers have developed various computational techniques using scalp electroencephalography (EEG), intracranial EEG, and other neuroimaging modalities, evolving from traditional signal processing methods (e.g., wavelet transforms and template matching) to advanced machine learning (ML, e.g., support vector machines and random forests) and deep learning (DL) algorithms (e.g., convolutional neural networks, recurrent neural networks, transformers, graph neural networks, and hybrid architectures). This review provides a detailed examination of epileptic seizure detection and prediction, covering the key aspects of signal processing, ML algorithms, and DL techniques applied to brainwave signals. We systematically categorized the techniques, analyzed key research trends, and identified critical challenges (e.g., data scarcity, model generalizability, and real-time processing). By highlighting the gaps in the literature, this review serves as a valuable resource for researchers and offers insights into future directions for improving the accuracy, interpretability, and clinical applicability of EEG-based seizure detection systems.</p>","PeriodicalId":29931,"journal":{"name":"Visual Computing for Industry Biomedicine and Art","volume":"8 1","pages":"27"},"PeriodicalIF":6.0,"publicationDate":"2025-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12696252/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145726481","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Body Cosmos 2.0: embodied biofeedback interface for dancing. Body Cosmos 2.0:舞蹈的具身生物反馈界面。
IF 6 4区 计算机科学 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-11-20 DOI: 10.1186/s42492-025-00207-9
Rem RunGu Lin, Koo Yongen Ke, Kang Zhang

This study presents Body Cosmos 2.0, an embodied biofeedback system with an interactive interface situated at the intersection of dance, human-computer interaction, and bio-art. Building on the authors' prior work, "Body Cosmos: An Immersive Experience Driven by Real-time Bio-data," the system presents the concept of a 'bio-body'-a dynamic digital embodiment of a dancer's internal state-generated in real-time through electroencephalography, heart rate sensors, motion tracking, and visualization techniques. Dancers interact with the system through three distinct experiences "VR embodiment," which enables them to experience their internal states from a first-person perspective; "dancing within your bio-body," which immerses them in their internal physiological and emotional states; and "dancing with your bio-body," which creates a bio-digital reflection for expressive development and experiential exploration. To evaluate the system's effectiveness, a workshop was conducted with 24 experienced dancers to assess its impact on self-awareness, creativity, and dance expressions. This integration of biodata with artistic expression transcends traditional neurofeedback and delves into the realm of embodied cognition. The study explores the concept, development, and application of "Body Cosmos 2.0," highlighting its potential to amplify self-awareness, augment performance, and expand the expressive and creative possibilities of dance.

本研究提出了身体宇宙2.0,一个具身的生物反馈系统,其交互界面位于舞蹈,人机交互和生物艺术的交叉点。在作者之前的工作“身体宇宙:实时生物数据驱动的沉浸式体验”的基础上,该系统提出了“生物身体”的概念——通过脑电图、心率传感器、运动跟踪和可视化技术实时生成的舞者内部状态的动态数字体现。舞者通过三种不同的体验与系统互动:“VR化身”,使他们能够从第一人称的角度体验他们的内部状态;“在你的生物体内跳舞”,让他们沉浸在内在的生理和情感状态中;“与你的生物身体共舞”,它创造了一种生物数字反射,用于表达发展和体验探索。为了评估该系统的有效性,我们与24名经验丰富的舞者进行了一次研讨会,以评估其对自我意识、创造力和舞蹈表达的影响。这种生物数据与艺术表达的结合超越了传统的神经反馈,深入到具身认知的领域。该研究探讨了“身体宇宙2.0”的概念、发展和应用,强调了它在增强自我意识、增强表演、扩大舞蹈表达和创造可能性方面的潜力。
{"title":"Body Cosmos 2.0: embodied biofeedback interface for dancing.","authors":"Rem RunGu Lin, Koo Yongen Ke, Kang Zhang","doi":"10.1186/s42492-025-00207-9","DOIUrl":"10.1186/s42492-025-00207-9","url":null,"abstract":"<p><p>This study presents Body Cosmos 2.0, an embodied biofeedback system with an interactive interface situated at the intersection of dance, human-computer interaction, and bio-art. Building on the authors' prior work, \"Body Cosmos: An Immersive Experience Driven by Real-time Bio-data,\" the system presents the concept of a 'bio-body'-a dynamic digital embodiment of a dancer's internal state-generated in real-time through electroencephalography, heart rate sensors, motion tracking, and visualization techniques. Dancers interact with the system through three distinct experiences \"VR embodiment,\" which enables them to experience their internal states from a first-person perspective; \"dancing within your bio-body,\" which immerses them in their internal physiological and emotional states; and \"dancing with your bio-body,\" which creates a bio-digital reflection for expressive development and experiential exploration. To evaluate the system's effectiveness, a workshop was conducted with 24 experienced dancers to assess its impact on self-awareness, creativity, and dance expressions. This integration of biodata with artistic expression transcends traditional neurofeedback and delves into the realm of embodied cognition. The study explores the concept, development, and application of \"Body Cosmos 2.0,\" highlighting its potential to amplify self-awareness, augment performance, and expand the expressive and creative possibilities of dance.</p>","PeriodicalId":29931,"journal":{"name":"Visual Computing for Industry Biomedicine and Art","volume":"8 1","pages":"26"},"PeriodicalIF":6.0,"publicationDate":"2025-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12634995/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145565441","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Applications of extended reality in pilot flight simulator training: a systematic review with meta-analysis. 扩展现实在飞行员飞行模拟器训练中的应用:系统回顾与元分析。
IF 6 4区 计算机科学 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-10-23 DOI: 10.1186/s42492-025-00206-w
Alexander Somerville, Keith Joiner, Timothy Lynar, Graham Wild

The use of extended reality (XR) spectrum technologies as substitutes to augment traditional simulators in pilot flight training has received significant interest in recent times. A systematic review was conducted to evaluate the efficacy of XR technologies for this purpose and better understand the motivating factors for this use. The systematic review followed the QUOROM framework (adapted for educational studies), screening 1237 candidate articles to 67 eligible for thematic analysis, with 5 of these also meeting meta-analysis criteria. Existing literature emphasizes the benefits of these technologies, particularly as a result of immersiveness and spatial awareness, enabling the application of more modern educational theories. Although the existing literature is concerned with much of the industry, there is a specific focus on general aviation and the more ab initio skills of flight. The results of the meta-analysis indicate improvements in pilot performance, with an overall meta-analytic effect size estimate of 0.884 (z = 2.248, P = 0.025), which is positive, statistically significant, and moderately strong. The findings of this review indicate support for the use and intention for the use of XR in pilot flight training simulators. However, multiple serious research gaps exist, such as the potential higher occurrence of simulator sickness and cybersickness, and a lack of robust research trials that examine transfer of training across the full pilot skill set and curricular contexts. This novel systematic review and meta-analysis represent a significant attempt to shape and direct better research to help to direct flourishing technological XR development in a time of increasing pilot shortages and aviation growth.

近年来,在飞行员飞行训练中使用扩展现实(XR)频谱技术作为传统模拟器的替代品已经引起了人们的极大兴趣。我们进行了一项系统评价,以评估XR技术在这方面的功效,并更好地了解这种使用的激励因素。系统评价遵循QUOROM框架(适用于教育研究),筛选1237篇候选文章,其中67篇符合主题分析标准,其中5篇也符合元分析标准。现有文献强调了这些技术的好处,特别是由于沉浸式和空间意识,使更多现代教育理论的应用成为可能。虽然现有的文献是有关行业的大部分,有一个特别的重点是通用航空和更多的从头开始的飞行技能。meta分析结果表明,飞行员绩效有所改善,总体meta分析效应量估计为0.884 (z = 2.248, P = 0.025),具有统计学显著性,且中等强度。本综述的研究结果表明支持XR在飞行员飞行训练模拟器中的使用和意图。然而,存在许多严重的研究空白,例如模拟器病和晕动病的发生率可能更高,以及缺乏检查整个飞行员技能和课程背景下培训转移的强有力的研究试验。这项新颖的系统综述和荟萃分析代表了一项重要的尝试,旨在塑造和指导更好的研究,以帮助指导在飞行员短缺和航空业增长日益严重的情况下蓬勃发展的技术XR发展。
{"title":"Applications of extended reality in pilot flight simulator training: a systematic review with meta-analysis.","authors":"Alexander Somerville, Keith Joiner, Timothy Lynar, Graham Wild","doi":"10.1186/s42492-025-00206-w","DOIUrl":"10.1186/s42492-025-00206-w","url":null,"abstract":"<p><p>The use of extended reality (XR) spectrum technologies as substitutes to augment traditional simulators in pilot flight training has received significant interest in recent times. A systematic review was conducted to evaluate the efficacy of XR technologies for this purpose and better understand the motivating factors for this use. The systematic review followed the QUOROM framework (adapted for educational studies), screening 1237 candidate articles to 67 eligible for thematic analysis, with 5 of these also meeting meta-analysis criteria. Existing literature emphasizes the benefits of these technologies, particularly as a result of immersiveness and spatial awareness, enabling the application of more modern educational theories. Although the existing literature is concerned with much of the industry, there is a specific focus on general aviation and the more ab initio skills of flight. The results of the meta-analysis indicate improvements in pilot performance, with an overall meta-analytic effect size estimate of 0.884 (z = 2.248, P = 0.025), which is positive, statistically significant, and moderately strong. The findings of this review indicate support for the use and intention for the use of XR in pilot flight training simulators. However, multiple serious research gaps exist, such as the potential higher occurrence of simulator sickness and cybersickness, and a lack of robust research trials that examine transfer of training across the full pilot skill set and curricular contexts. This novel systematic review and meta-analysis represent a significant attempt to shape and direct better research to help to direct flourishing technological XR development in a time of increasing pilot shortages and aviation growth.</p>","PeriodicalId":29931,"journal":{"name":"Visual Computing for Industry Biomedicine and Art","volume":"8 1","pages":"25"},"PeriodicalIF":6.0,"publicationDate":"2025-10-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12546163/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145348911","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
KaiBiLi: gesture-based immersive virtual reality ceremony for traditional Chinese cultural activities. KaiBiLi:基于手势的沉浸式虚拟现实仪式,用于中国传统文化活动。
IF 6 4区 计算机科学 Q2 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-10-02 DOI: 10.1186/s42492-025-00205-x
Yiping Wu, Yue Li, Eugene Ch'ng, Jiaxin Gao, Tao Hong

Gesture-based interactions in a virtual reality (VR) setting can enhance our experience of traditional practices as part of preserving and communicating heritage. Cultural experiences embodied within VR environments are suggested to be an effective approach for experiencing intangible cultural heritage. Ceremonies, rituals, and related ancestral enactments are important for preserving cultural heritage. Kāi Bǐ Lǐ, also known as the First Writing Ceremony, is traditionally held for Chinese children before their first year of elementary school. However, gesture-based immersive VR for learning this tradition is new, and have not been developed within the community. This study focused on how users experienced learning cultural practices using gesture-based interactive VR across different age groups and hardware platforms. We first conducted an experiment with 60 participants (30 young adults and 30 children) using the First Writing Ceremony as a case study in which gestural interactions were elicited, designed, implemented, and evaluated. The study showed significant differences in play time and presence between the head-mounted display VR and desktop VR. In addition, children were less likely to experience fatigue than young adults. Following this, we conducted another study after eight months to investigate the VR systems' long-term learning effectiveness. This showed that children outperformed young adults in demonstrating greater knowledge retention. Our results and findings contribute to the design of gesture-based VR for different age groups across different platforms for experiencing, learning, and practicing cultural activities.

虚拟现实(VR)环境中基于手势的互动可以增强我们对传统习俗的体验,作为保护和传播遗产的一部分。在VR环境中体现文化体验是体验非物质文化遗产的有效途径。仪式、仪式和相关的祖传法令对保护文化遗产很重要。Kāi b / l /,也被称为初笔礼,传统上是为中国孩子在小学一年级之前举行的。然而,用于学习这一传统的基于手势的沉浸式VR是新的,并且尚未在社区中开发出来。这项研究的重点是用户如何在不同年龄组和硬件平台上使用基于手势的交互式VR体验学习文化实践。我们首先对60名参与者(30名年轻人和30名儿童)进行了实验,以“第一次书写仪式”为例研究手势互动的引发、设计、实施和评估。研究显示,头戴式虚拟现实和桌面虚拟现实在游戏时间和存在感上存在显著差异。此外,儿童比年轻人更不容易感到疲劳。在此之后,我们在8个月后进行了另一项研究,以调查VR系统的长期学习效果。这表明,儿童在知识记忆方面比年轻人表现得更好。我们的研究结果和发现有助于设计基于手势的VR,适用于不同年龄段、不同平台的体验、学习和实践文化活动。
{"title":"KaiBiLi: gesture-based immersive virtual reality ceremony for traditional Chinese cultural activities.","authors":"Yiping Wu, Yue Li, Eugene Ch'ng, Jiaxin Gao, Tao Hong","doi":"10.1186/s42492-025-00205-x","DOIUrl":"10.1186/s42492-025-00205-x","url":null,"abstract":"<p><p>Gesture-based interactions in a virtual reality (VR) setting can enhance our experience of traditional practices as part of preserving and communicating heritage. Cultural experiences embodied within VR environments are suggested to be an effective approach for experiencing intangible cultural heritage. Ceremonies, rituals, and related ancestral enactments are important for preserving cultural heritage. Kāi Bǐ Lǐ, also known as the First Writing Ceremony, is traditionally held for Chinese children before their first year of elementary school. However, gesture-based immersive VR for learning this tradition is new, and have not been developed within the community. This study focused on how users experienced learning cultural practices using gesture-based interactive VR across different age groups and hardware platforms. We first conducted an experiment with 60 participants (30 young adults and 30 children) using the First Writing Ceremony as a case study in which gestural interactions were elicited, designed, implemented, and evaluated. The study showed significant differences in play time and presence between the head-mounted display VR and desktop VR. In addition, children were less likely to experience fatigue than young adults. Following this, we conducted another study after eight months to investigate the VR systems' long-term learning effectiveness. This showed that children outperformed young adults in demonstrating greater knowledge retention. Our results and findings contribute to the design of gesture-based VR for different age groups across different platforms for experiencing, learning, and practicing cultural activities.</p>","PeriodicalId":29931,"journal":{"name":"Visual Computing for Industry Biomedicine and Art","volume":"8 1","pages":"24"},"PeriodicalIF":6.0,"publicationDate":"2025-10-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12491145/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145207799","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Visual Computing for Industry Biomedicine and Art
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1