首页 > 最新文献

Advances in visual computing : ... international symposium, ISVC ... : proceedings. International Symposium on Visual Computing最新文献

英文 中文
Biomimetic Oculomotor Control with Spiking Neural Networks 基于脉冲神经网络的仿生眼动控制
T. Saquib, Demetri Terzopoulos
{"title":"Biomimetic Oculomotor Control with Spiking Neural Networks","authors":"T. Saquib, Demetri Terzopoulos","doi":"10.1007/978-3-031-20716-7_2","DOIUrl":"https://doi.org/10.1007/978-3-031-20716-7_2","url":null,"abstract":"","PeriodicalId":91444,"journal":{"name":"Advances in visual computing : ... international symposium, ISVC ... : proceedings. International Symposium on Visual Computing","volume":"1 1","pages":"1-14"},"PeriodicalIF":0.0,"publicationDate":"2023-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89744266","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Saliency Can Be All You Need In Contrastive Self-Supervised Learning 在对比自我监督学习中,显著性是你所需要的
Veysel Kocaman, O. M. Shir, Thomas Bäck, A. Belbachir
We propose an augmentation policy for Contrastive Self-Supervised Learning (SSL) in the form of an already established Salient Image Segmentation technique entitled Global Contrast based Salient Region Detection. This detection technique, which had been devised for unrelated Computer Vision tasks, was empirically observed to play the role of an augmentation facilitator within the SSL protocol. This observation is rooted in our practical attempts to learn, by SSL-fashion, aerial imagery of solar panels, which exhibit challenging bound-ary patterns. Upon the successful integration of this technique on our problem domain, we formulated a generalized procedure and conducted a comprehensive, systematic performance assessment with various Contrastive SSL algorithms subject to standard augmentation techniques. This evaluation, which was conducted across multiple datasets, indicated that the proposed technique indeed contributes to SSL. We hypothesize whether salient image segmentation may suffice as the only augmentation policy in Contrastive SSL when treating downstream segmentation tasks.
我们提出了一种增强对比自监督学习(SSL)的策略,其形式是一种已经建立的显著图像分割技术,称为基于全局对比度的显著区域检测。这种检测技术是为不相关的计算机视觉任务设计的,根据经验观察,它在SSL协议中发挥了增强促进者的作用。这种观察根植于我们实际尝试学习,通过ssl时尚,太阳能电池板的航空图像,展示具有挑战性的边界模式。在成功地将该技术集成到我们的问题领域之后,我们制定了一个一般化的过程,并使用符合标准增强技术的各种对比SSL算法进行了全面、系统的性能评估。这个跨多个数据集进行的评估表明,所提议的技术确实有助于SSL。我们假设在处理下游分割任务时,显著图像分割是否足以作为对比度SSL中唯一的增强策略。
{"title":"Saliency Can Be All You Need In Contrastive Self-Supervised Learning","authors":"Veysel Kocaman, O. M. Shir, Thomas Bäck, A. Belbachir","doi":"10.48550/arXiv.2210.16776","DOIUrl":"https://doi.org/10.48550/arXiv.2210.16776","url":null,"abstract":"We propose an augmentation policy for Contrastive Self-Supervised Learning (SSL) in the form of an already established Salient Image Segmentation technique entitled Global Contrast based Salient Region Detection. This detection technique, which had been devised for unrelated Computer Vision tasks, was empirically observed to play the role of an augmentation facilitator within the SSL protocol. This observation is rooted in our practical attempts to learn, by SSL-fashion, aerial imagery of solar panels, which exhibit challenging bound-ary patterns. Upon the successful integration of this technique on our problem domain, we formulated a generalized procedure and conducted a comprehensive, systematic performance assessment with various Contrastive SSL algorithms subject to standard augmentation techniques. This evaluation, which was conducted across multiple datasets, indicated that the proposed technique indeed contributes to SSL. We hypothesize whether salient image segmentation may suffice as the only augmentation policy in Contrastive SSL when treating downstream segmentation tasks.","PeriodicalId":91444,"journal":{"name":"Advances in visual computing : ... international symposium, ISVC ... : proceedings. International Symposium on Visual Computing","volume":"207 1","pages":"119-140"},"PeriodicalIF":0.0,"publicationDate":"2022-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78774374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A DirectX-Based DICOM Viewer for Multi-User Surgical Planning in Augmented Reality 基于directx的增强现实多用户手术计划DICOM查看器
Menghe Zhang, Weichen Liu, Nadir Weibel, J. Schulze
Preoperative medical imaging is an essential part of surgical planning. The data from medical imaging devices, such as CT and MRI scanners, consist of stacks of 2D images in DICOM format. Conversely, advances in 3D data visualization provide further information by assembling cross-sections into 3D volumetric datasets. As Microsoft unveiled the HoloLens 2 (HL2), which is considered one of the best Mixed Reality (XR) headsets in the market, it promised to enhance visualization in 3D by providing an immersive experience to users. This paper introduces a prototype holographic XR DICOM Viewer for the 3D visualization of DICOM image sets on HL2 for surgical planning. We first developed a standalone graphical C++ engine using the native DirectX11 API and HLSL shaders. Based on that, the prototype further applies the OpenXR API for potential deployment on a wide range of devices from vendors across the XR spectrum. With native access to the device, our prototype unravels the limitation of hardware capabilities on HL2 for 3D volume rendering and interaction. Moreover, smartphones can act as input devices to provide another user interaction method by connecting to our server. In this paper, we present a holographic DICOM viewer for the HoloLens 2 and contribute (i) a prototype that renders the DICOM image stacks in real-time on HL2, (ii) three types of user interactions in XR, and (iii) a preliminary qualitative evaluation of our prototype.
术前医学影像是手术计划的重要组成部分。来自医学成像设备(如CT和MRI扫描仪)的数据由DICOM格式的2D图像堆栈组成。相反,3D数据可视化的进步通过将横截面组装成3D体积数据集提供了进一步的信息。微软推出的HoloLens 2 (HL2)被认为是市场上最好的混合现实(XR)头显之一,它承诺通过为用户提供身临其境的体验来增强3D可视化。本文介绍了一种用于HL2上DICOM图像集三维可视化的全息XR DICOM查看器原型。我们首先使用本地DirectX11 API和hsl着色器开发了一个独立的图形化c++引擎。在此基础上,原型进一步应用OpenXR API,以便在XR范围内的供应商提供的各种设备上进行潜在部署。通过本机访问设备,我们的原型解决了HL2上用于3D体渲染和交互的硬件功能的限制。此外,智能手机可以作为输入设备,通过连接到我们的服务器提供另一种用户交互方式。在本文中,我们为HoloLens 2提供了一个全息DICOM查看器,并提供了(i)一个在HL2上实时呈现DICOM图像堆栈的原型,(ii) XR中的三种类型的用户交互,以及(iii)我们的原型的初步定性评估。
{"title":"A DirectX-Based DICOM Viewer for Multi-User Surgical Planning in Augmented Reality","authors":"Menghe Zhang, Weichen Liu, Nadir Weibel, J. Schulze","doi":"10.48550/arXiv.2210.14349","DOIUrl":"https://doi.org/10.48550/arXiv.2210.14349","url":null,"abstract":"Preoperative medical imaging is an essential part of surgical planning. The data from medical imaging devices, such as CT and MRI scanners, consist of stacks of 2D images in DICOM format. Conversely, advances in 3D data visualization provide further information by assembling cross-sections into 3D volumetric datasets. As Microsoft unveiled the HoloLens 2 (HL2), which is considered one of the best Mixed Reality (XR) headsets in the market, it promised to enhance visualization in 3D by providing an immersive experience to users. This paper introduces a prototype holographic XR DICOM Viewer for the 3D visualization of DICOM image sets on HL2 for surgical planning. We first developed a standalone graphical C++ engine using the native DirectX11 API and HLSL shaders. Based on that, the prototype further applies the OpenXR API for potential deployment on a wide range of devices from vendors across the XR spectrum. With native access to the device, our prototype unravels the limitation of hardware capabilities on HL2 for 3D volume rendering and interaction. Moreover, smartphones can act as input devices to provide another user interaction method by connecting to our server. In this paper, we present a holographic DICOM viewer for the HoloLens 2 and contribute (i) a prototype that renders the DICOM image stacks in real-time on HL2, (ii) three types of user interactions in XR, and (iii) a preliminary qualitative evaluation of our prototype.","PeriodicalId":91444,"journal":{"name":"Advances in visual computing : ... international symposium, ISVC ... : proceedings. International Symposium on Visual Computing","volume":"39 1","pages":"217-228"},"PeriodicalIF":0.0,"publicationDate":"2022-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87483317","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ConnectedUNets++: Mass Segmentation from Whole Mammographic Images connectedunet++:从整个乳房x线摄影图像的质量分割
Prithul Sarker, Sushmita Sarker, G. Bebis, A. Tavakkoli
Deep learning has made a breakthrough in medical image segmentation in recent years due to its ability to extract high-level features without the need for prior knowledge. In this context, U-Net is one of the most advanced medical image segmentation models, with promising results in mammography. Despite its excellent overall performance in segmenting multimodal medical images, the traditional U-Net structure appears to be inadequate in various ways. There are certain U-Net design modifications, such as MultiResUNet, Connected-UNets, and AU-Net, that have improved overall performance in areas where the conventional U-Net architecture appears to be deficient. Following the success of UNet and its variants, we have presented two enhanced versions of the Connected-UNets architecture: ConnectedUNets+ and ConnectedUNets++. In ConnectedUNets+, we have replaced the simple skip connections of Connected-UNets architecture with residual skip connections, while in ConnectedUNets++, we have modified the encoder-decoder structure along with employing residual skip connections. We have evaluated our proposed architectures on two publicly available datasets, the Curated Breast Imaging Subset of Digital Database for Screening Mammography (CBIS-DDSM) and INbreast.
近年来,深度学习在医学图像分割方面取得了突破性进展,因为它能够在不需要先验知识的情况下提取高级特征。在此背景下,U-Net是最先进的医学图像分割模型之一,在乳房x光检查中有很好的效果。尽管传统的U-Net结构在多模态医学图像分割方面具有优异的综合性能,但在许多方面存在不足。有一些U-Net的设计修改,如MultiResUNet、connected - unet和AU-Net,在传统U-Net架构不足的地方提高了整体性能。随着UNet及其变体的成功,我们提出了两个增强版本的Connected-UNets架构:connectedunets++和connectedunets++。在ConnectedUNets+中,我们用剩余的跳过连接取代了ConnectedUNets架构的简单跳过连接,而在ConnectedUNets++中,我们修改了编码器-解码器结构,并使用剩余的跳过连接。我们在两个公开可用的数据集上评估了我们提出的架构,这两个数据集是乳腺造影筛查数字数据库(CBIS-DDSM)和INbreast。
{"title":"ConnectedUNets++: Mass Segmentation from Whole Mammographic Images","authors":"Prithul Sarker, Sushmita Sarker, G. Bebis, A. Tavakkoli","doi":"10.48550/arXiv.2210.13668","DOIUrl":"https://doi.org/10.48550/arXiv.2210.13668","url":null,"abstract":"Deep learning has made a breakthrough in medical image segmentation in recent years due to its ability to extract high-level features without the need for prior knowledge. In this context, U-Net is one of the most advanced medical image segmentation models, with promising results in mammography. Despite its excellent overall performance in segmenting multimodal medical images, the traditional U-Net structure appears to be inadequate in various ways. There are certain U-Net design modifications, such as MultiResUNet, Connected-UNets, and AU-Net, that have improved overall performance in areas where the conventional U-Net architecture appears to be deficient. Following the success of UNet and its variants, we have presented two enhanced versions of the Connected-UNets architecture: ConnectedUNets+ and ConnectedUNets++. In ConnectedUNets+, we have replaced the simple skip connections of Connected-UNets architecture with residual skip connections, while in ConnectedUNets++, we have modified the encoder-decoder structure along with employing residual skip connections. We have evaluated our proposed architectures on two publicly available datasets, the Curated Breast Imaging Subset of Digital Database for Screening Mammography (CBIS-DDSM) and INbreast.","PeriodicalId":91444,"journal":{"name":"Advances in visual computing : ... international symposium, ISVC ... : proceedings. International Symposium on Visual Computing","volume":"11 1","pages":"419-430"},"PeriodicalIF":0.0,"publicationDate":"2022-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81430091","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pruning-based Topology Refinement of 3D Mesh using a 2D Alpha Mask 基于剪枝的二维Alpha蒙版三维网格拓扑优化
Ga'etan Landreau, M. Tamaazousti
Image-based 3D reconstruction has increasingly stunning results over the past few years with the latest improvements in computer vision and graphics. Geometry and topology are two fundamental concepts when dealing with 3D mesh structures. But the latest often remains a side issue in the 3D mesh-based reconstruction literature. Indeed, performing per-vertex elementary displacements over a 3D sphere mesh only impacts its geometry and leaves the topological structure unchanged and fixed. Whereas few attempts propose to update the geometry and the topology, all need to lean on costly 3D ground-truth to determine the faces/edges to prune. We present in this work a method that aims to refine the topology of any 3D mesh through a face-pruning strategy that extensively relies upon 2D alpha masks and camera pose information. Our solution leverages a differentiable renderer that renders each face as a 2D soft map. Its pixel intensity reflects the probability of being covered during the rendering process by such a face. Based on the 2D soft-masks available, our method is thus able to quickly highlight all the incorrectly rendered faces for a given viewpoint. Because our module is agnostic to the network that produces the 3D mesh, it can be easily plugged into any self-supervised image-based (either synthetic or natural) 3D reconstruction pipeline to get complex meshes with a non-spherical topology.
在过去的几年里,随着计算机视觉和图形学的最新进步,基于图像的3D重建取得了越来越惊人的结果。几何和拓扑是处理三维网格结构时的两个基本概念。但最新的通常仍然是一个侧面问题,在3D网格为基础的重建文献。实际上,在3D球体网格上执行逐顶点基本位移只会影响其几何形状,而拓扑结构不变且固定。然而,很少有人尝试更新几何形状和拓扑结构,所有这些都需要依靠昂贵的3D地面真相来确定要修剪的面/边。我们在这项工作中提出了一种方法,旨在通过广泛依赖于2D alpha蒙版和相机姿态信息的面部修剪策略来细化任何3D网格的拓扑结构。我们的解决方案利用了一个可微分渲染器,将每个面渲染为2D软地图。其像素强度反映了在渲染过程中被该人脸覆盖的概率。基于可用的2D软蒙版,我们的方法因此能够快速突出显示给定视点的所有不正确渲染的面部。因为我们的模块与产生3D网格的网络无关,所以它可以很容易地插入任何基于自监督图像的(合成或自然)3D重建管道中,以获得具有非球面拓扑结构的复杂网格。
{"title":"Pruning-based Topology Refinement of 3D Mesh using a 2D Alpha Mask","authors":"Ga'etan Landreau, M. Tamaazousti","doi":"10.48550/arXiv.2210.09148","DOIUrl":"https://doi.org/10.48550/arXiv.2210.09148","url":null,"abstract":"Image-based 3D reconstruction has increasingly stunning results over the past few years with the latest improvements in computer vision and graphics. Geometry and topology are two fundamental concepts when dealing with 3D mesh structures. But the latest often remains a side issue in the 3D mesh-based reconstruction literature. Indeed, performing per-vertex elementary displacements over a 3D sphere mesh only impacts its geometry and leaves the topological structure unchanged and fixed. Whereas few attempts propose to update the geometry and the topology, all need to lean on costly 3D ground-truth to determine the faces/edges to prune. We present in this work a method that aims to refine the topology of any 3D mesh through a face-pruning strategy that extensively relies upon 2D alpha masks and camera pose information. Our solution leverages a differentiable renderer that renders each face as a 2D soft map. Its pixel intensity reflects the probability of being covered during the rendering process by such a face. Based on the 2D soft-masks available, our method is thus able to quickly highlight all the incorrectly rendered faces for a given viewpoint. Because our module is agnostic to the network that produces the 3D mesh, it can be easily plugged into any self-supervised image-based (either synthetic or natural) 3D reconstruction pipeline to get complex meshes with a non-spherical topology.","PeriodicalId":91444,"journal":{"name":"Advances in visual computing : ... international symposium, ISVC ... : proceedings. International Symposium on Visual Computing","volume":"14 1","pages":"404-415"},"PeriodicalIF":0.0,"publicationDate":"2022-10-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88416540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Learning based Super-Resolution for Medical Volume Visualization with Direct Volume Rendering 基于深度学习的超分辨率医学体可视化直接体绘制
S. Devkota, S. Pattanaik
Modern-day display systems demand high-quality rendering. However, rendering at higher resolution requires a large number of data samples and is computationally expensive. Recent advances in deep learning-based image and video super-resolution techniques motivate us to investigate such networks for high-fidelity upscaling of frames rendered at a lower resolution to a higher resolution. While our work focuses on super-resolution of medical volume visualization performed with direct volume rendering, it is also applicable for volume visualization with other rendering techniques. We propose a learning-based technique where our proposed system uses color information along with other supplementary features gathered from our volume renderer to learn efficient upscaling of a low-resolution rendering to a higher-resolution space. Furthermore, to improve temporal stability, we also implement the temporal reprojection technique for accumulating history samples in volumetric rendering.
现代显示系统需要高质量的渲染。然而,在更高的分辨率下渲染需要大量的数据样本,并且计算成本很高。基于深度学习的图像和视频超分辨率技术的最新进展促使我们研究这种网络,以高保真地将低分辨率渲染的帧升级到高分辨率。虽然我们的工作重点是使用直接体绘制进行超分辨率的医学体可视化,但它也适用于使用其他渲染技术进行体可视化。我们提出了一种基于学习的技术,其中我们提出的系统使用颜色信息以及从我们的体渲染器收集的其他补充特征来学习将低分辨率渲染有效地升级到更高分辨率的空间。此外,为了提高时间稳定性,我们还实现了在体绘制中积累历史样本的时间重投影技术。
{"title":"Deep Learning based Super-Resolution for Medical Volume Visualization with Direct Volume Rendering","authors":"S. Devkota, S. Pattanaik","doi":"10.48550/arXiv.2210.08080","DOIUrl":"https://doi.org/10.48550/arXiv.2210.08080","url":null,"abstract":"Modern-day display systems demand high-quality rendering. However, rendering at higher resolution requires a large number of data samples and is computationally expensive. Recent advances in deep learning-based image and video super-resolution techniques motivate us to investigate such networks for high-fidelity upscaling of frames rendered at a lower resolution to a higher resolution. While our work focuses on super-resolution of medical volume visualization performed with direct volume rendering, it is also applicable for volume visualization with other rendering techniques. We propose a learning-based technique where our proposed system uses color information along with other supplementary features gathered from our volume renderer to learn efficient upscaling of a low-resolution rendering to a higher-resolution space. Furthermore, to improve temporal stability, we also implement the temporal reprojection technique for accumulating history samples in volumetric rendering.","PeriodicalId":91444,"journal":{"name":"Advances in visual computing : ... international symposium, ISVC ... : proceedings. International Symposium on Visual Computing","volume":"89 1","pages":"103-114"},"PeriodicalIF":0.0,"publicationDate":"2022-10-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80252632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Virtual-Reality based Vestibular Ocular Motor Screening for Concussion Detection using Machine-Learning 基于虚拟现实的前庭眼运动筛查用于脑震荡检测的机器学习
Khondker Fariha Hossain, Sharif Amit Kamran, Prithul Sarker, Philip Pavilionis, I. Adhanom, N. Murray, A. Tavakkoli
. Sport-related concussion (SRC) depends on sensory information from visual, vestibular, and somatosensory systems. At the same time, the current clinical administration of Vestibular/Ocular Motor Screening (VOMS) is subjective and deviates among administrators. Therefore, for the assessment and manage-ment of concussion detection, standardization is required to lower the risk of injury and increase the validation among clinicians. With the advancement of technology, virtual reality (VR) can be utilized to advance the standardization of the VOMS, increasing the accuracy of testing administration and decreasing overall false positive rates. In this paper, we experimented with multiple machine learning methods to detect SRC on VR-generated data using VOMS. In our observation, the data generated from VR for smooth pursuit (SP) and the Visual Motion Sensitivity (VMS) tests are highly reliable for concussion detection. Furthermore, we train and evaluate these models, both qualitatively and quan-titatively. Our findings show these models can reach high true-positive-rates of around 99.9 percent of symptom provocation on the VR stimuli-based VOMS vs. current clinical manual VOMS.
。运动相关脑震荡(SRC)依赖于视觉、前庭和体感系统的感觉信息。同时,目前临床对前庭/眼运动筛查(VOMS)的管理是主观的,并且在管理人员之间存在偏差。因此,对于脑震荡检测的评估和管理,需要标准化,以降低损伤的风险,并增加临床医生的验证。随着技术的进步,虚拟现实技术可以促进VOMS的标准化,提高检测管理的准确性,降低总体假阳性率。在本文中,我们使用VOMS实验了多种机器学习方法来检测vr生成数据上的SRC。在我们的观察中,由VR生成的平滑追踪(SP)和视觉运动灵敏度(VMS)测试数据对于脑震荡检测是高度可靠的。此外,我们对这些模型进行了定性和定量的训练和评估。我们的研究结果表明,与目前的临床手动VOMS相比,这些模型在基于VR刺激的VOMS上可以达到99.9%左右的高真阳性率。
{"title":"Virtual-Reality based Vestibular Ocular Motor Screening for Concussion Detection using Machine-Learning","authors":"Khondker Fariha Hossain, Sharif Amit Kamran, Prithul Sarker, Philip Pavilionis, I. Adhanom, N. Murray, A. Tavakkoli","doi":"10.48550/arXiv.2210.09295","DOIUrl":"https://doi.org/10.48550/arXiv.2210.09295","url":null,"abstract":". Sport-related concussion (SRC) depends on sensory information from visual, vestibular, and somatosensory systems. At the same time, the current clinical administration of Vestibular/Ocular Motor Screening (VOMS) is subjective and deviates among administrators. Therefore, for the assessment and manage-ment of concussion detection, standardization is required to lower the risk of injury and increase the validation among clinicians. With the advancement of technology, virtual reality (VR) can be utilized to advance the standardization of the VOMS, increasing the accuracy of testing administration and decreasing overall false positive rates. In this paper, we experimented with multiple machine learning methods to detect SRC on VR-generated data using VOMS. In our observation, the data generated from VR for smooth pursuit (SP) and the Visual Motion Sensitivity (VMS) tests are highly reliable for concussion detection. Furthermore, we train and evaluate these models, both qualitatively and quan-titatively. Our findings show these models can reach high true-positive-rates of around 99.9 percent of symptom provocation on the VR stimuli-based VOMS vs. current clinical manual VOMS.","PeriodicalId":91444,"journal":{"name":"Advances in visual computing : ... international symposium, ISVC ... : proceedings. International Symposium on Visual Computing","volume":"30 1","pages":"229-241"},"PeriodicalIF":0.0,"publicationDate":"2022-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83421033","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Game Theoretical vulnerability analysis of Adversarial Attack 对抗性攻击的博弈脆弱性分析
Khondker Fariha Hossain, A. Tavakkoli, S. Sengupta
In recent times deep learning has been widely used for automating various security tasks in Cyber Domains. However, adversaries manipulate data in many situations and diminish the deployed deep learning model's accuracy. One notable example is fooling CAPTCHA data to access the CAPTCHA-based Classifier leading to the critical system being vulnerable to cybersecurity attacks. To alleviate this, we propose a computational framework of game theory to analyze the CAPTCHA-based Classifier's vulnerability, strategy, and outcomes by forming a simultaneous two-player game. We apply the Fast Gradient Symbol Method (FGSM) and One Pixel Attack on CAPTCHA Data to imitate real-life scenarios of possible cyber-attack. Subsequently, to interpret this scenario from a Game theoretical perspective, we represent the interaction in the Stackelberg Game in Kuhn tree to study players' possible behaviors and actions by applying our Classifier's actual predicted values. Thus, we interpret potential attacks in deep learning applications while representing viable defense strategies in the game theory prospect.
近年来,深度学习已被广泛用于网络领域各种安全任务的自动化。然而,攻击者在许多情况下操纵数据,降低了部署的深度学习模型的准确性。一个值得注意的例子是欺骗CAPTCHA数据来访问基于CAPTCHA的分类器,导致关键系统容易受到网络安全攻击。为了缓解这一问题,我们提出了一个博弈论的计算框架,通过形成一个同步的双人游戏来分析基于captcha的分类器的漏洞、策略和结果。我们应用快速梯度符号法(FGSM)和对验证码数据的一像素攻击来模拟可能的网络攻击的现实场景。随后,为了从博弈论的角度解释这一场景,我们将Stackelberg博弈中的相互作用表示为库恩树,通过应用我们的分类器的实际预测值来研究参与者可能的行为和行动。因此,我们解释了深度学习应用中潜在的攻击,同时代表了博弈论前景中可行的防御策略。
{"title":"A Game Theoretical vulnerability analysis of Adversarial Attack","authors":"Khondker Fariha Hossain, A. Tavakkoli, S. Sengupta","doi":"10.48550/arXiv.2210.06670","DOIUrl":"https://doi.org/10.48550/arXiv.2210.06670","url":null,"abstract":"In recent times deep learning has been widely used for automating various security tasks in Cyber Domains. However, adversaries manipulate data in many situations and diminish the deployed deep learning model's accuracy. One notable example is fooling CAPTCHA data to access the CAPTCHA-based Classifier leading to the critical system being vulnerable to cybersecurity attacks. To alleviate this, we propose a computational framework of game theory to analyze the CAPTCHA-based Classifier's vulnerability, strategy, and outcomes by forming a simultaneous two-player game. We apply the Fast Gradient Symbol Method (FGSM) and One Pixel Attack on CAPTCHA Data to imitate real-life scenarios of possible cyber-attack. Subsequently, to interpret this scenario from a Game theoretical perspective, we represent the interaction in the Stackelberg Game in Kuhn tree to study players' possible behaviors and actions by applying our Classifier's actual predicted values. Thus, we interpret potential attacks in deep learning applications while representing viable defense strategies in the game theory prospect.","PeriodicalId":91444,"journal":{"name":"Advances in visual computing : ... international symposium, ISVC ... : proceedings. International Symposium on Visual Computing","volume":"3 1","pages":"369-380"},"PeriodicalIF":0.0,"publicationDate":"2022-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88910089","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
VR-SFT: Reproducing Swinging Flashlight Test in Virtual Reality to Detect Relative Afferent Pupillary Defect VR-SFT:在虚拟现实中再现摆动手电筒测试以检测瞳孔相对传入缺陷
Prithul Sarker, Nasif Zaman, A. Tavakkoli
The relative afferent asymmetry between two eyes can be diagnosed using swinging flashlight test, also known as the alternating light test. This remains one of the most used clinical tests to this day. Despite the swinging flashlight test's straightforward approach, a number of factors can add variability into the clinical methodology and reduce the measurement's validity and reliability. This includes small and poorly responsive pupils, dark iris, anisocoria, uneven illumination in both eyes. Due to these limitations, the true condition of relative afferent asymmetry may create confusion and various observers may quantify the relative afferent pupillary defect differently. Consequently, the results of the swinging flashlight test are subjective and ambiguous. In order to eliminate the limitations of traditional swinging flashlight test and introduce objectivity, we propose a novel approach to the swinging flashlight exam, VR-SFT, by making use of virtual reality (VR). We suggest that the clinical records of the subjects and the results of VR-SFT are comparable. In this paper, we describe how we exploit the features of immersive VR experience to create a reliable and objective swinging flashlight test.
两只眼睛之间的相对传入不对称可以用摆动手电筒测试来诊断,也被称为交变光测试。直到今天,这仍然是最常用的临床试验之一。尽管摆动手电筒测试的方法简单明了,但许多因素会增加临床方法的可变性,从而降低测量的有效性和可靠性。这包括瞳孔小且反应差,虹膜暗,五光十色,双眼光照不均匀。由于这些限制,相对传入不对称的真实情况可能会造成混淆,不同的观察者可能会以不同的方式量化相对传入瞳孔缺陷。因此,摇摆手电筒测试的结果是主观的和模糊的。为了消除传统晃手电筒测试的局限性,引入客观性,我们提出了一种利用虚拟现实技术进行晃手电筒测试的新方法——VR- sft。我们认为受试者的临床记录与VR-SFT的结果具有可比性。在本文中,我们描述了我们如何利用沉浸式VR体验的特点来创建一个可靠和客观的摇摆手电筒测试。
{"title":"VR-SFT: Reproducing Swinging Flashlight Test in Virtual Reality to Detect Relative Afferent Pupillary Defect","authors":"Prithul Sarker, Nasif Zaman, A. Tavakkoli","doi":"10.48550/arXiv.2210.06474","DOIUrl":"https://doi.org/10.48550/arXiv.2210.06474","url":null,"abstract":"The relative afferent asymmetry between two eyes can be diagnosed using swinging flashlight test, also known as the alternating light test. This remains one of the most used clinical tests to this day. Despite the swinging flashlight test's straightforward approach, a number of factors can add variability into the clinical methodology and reduce the measurement's validity and reliability. This includes small and poorly responsive pupils, dark iris, anisocoria, uneven illumination in both eyes. Due to these limitations, the true condition of relative afferent asymmetry may create confusion and various observers may quantify the relative afferent pupillary defect differently. Consequently, the results of the swinging flashlight test are subjective and ambiguous. In order to eliminate the limitations of traditional swinging flashlight test and introduce objectivity, we propose a novel approach to the swinging flashlight exam, VR-SFT, by making use of virtual reality (VR). We suggest that the clinical records of the subjects and the results of VR-SFT are comparable. In this paper, we describe how we exploit the features of immersive VR experience to create a reliable and objective swinging flashlight test.","PeriodicalId":91444,"journal":{"name":"Advances in visual computing : ... international symposium, ISVC ... : proceedings. International Symposium on Visual Computing","volume":"20 1","pages":"193-204"},"PeriodicalIF":0.0,"publicationDate":"2022-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85268442","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Deep Labeling of fMRI Brain Networks Using Cloud Based Processing 基于云处理的fMRI脑网络深度标记
Sejal Ghate, Alberto Santamaría-Pang, I. Tarapov, H. Sair, Craig K. Jones
Resting state fMRI is an imaging modality which reveals brain activity localization through signal changes, in what is known as Resting State Networks (RSNs). This technique is gaining popularity in neurosurgical pre-planning to visualize the functional regions and assess regional activity. Labeling of rs-fMRI networks require subject-matter expertise and is time consuming, creating a need for an automated classification algorithm. While the impact of AI in medical diagnosis has shown great progress; deploying and maintaining these in a clinical setting is an unmet need. We propose an end-to-end reproducible pipeline which incorporates image processing of rs-fMRI in a cloud-based workflow while using deep learning to automate the classification of RSNs. We have architected a reproducible Azure Machine Learning cloud-based medical imaging concept pipeline for fMRI analysis integrating the popular FMRIB Software Library (FSL) toolkit. To demonstrate a clinical application using a large dataset, we compare three neural network architectures for classification of deeper RSNs derived from processed rs-fMRI. The three algorithms are: an MLP, a 2D projection-based CNN, and a fully 3D CNN classification networks. Each of the net-works was trained on the rs-fMRI back-projected independent components giving>98% accuracy for each classification method.
静息状态fMRI是一种通过信号变化揭示大脑活动定位的成像方式,即所谓的静息状态网络(RSNs)。这项技术在神经外科术前规划中越来越受欢迎,用于可视化功能区域和评估区域活动。标记rs-fMRI网络需要主题专业知识,并且耗时,因此需要自动分类算法。虽然人工智能在医疗诊断方面的影响取得了很大进展;在临床环境中部署和维护这些设备是一个未满足的需求。我们提出了一个端到端的可重复管道,该管道将rs-fMRI的图像处理整合到基于云的工作流中,同时使用深度学习来自动分类rsn。我们已经为fMRI分析构建了一个可重复的基于Azure机器学习云的医学成像概念管道,集成了流行的FMRIB软件库(FSL)工具包。为了演示使用大型数据集的临床应用,我们比较了三种神经网络架构,用于分类来自处理后的rs-fMRI的更深rsn。这三种算法分别是:MLP、基于2D投影的CNN和全3D CNN分类网络。每个网络都在rs-fMRI反向投影的独立分量上进行训练,每种分类方法的准确率都大于98%。
{"title":"Deep Labeling of fMRI Brain Networks Using Cloud Based Processing","authors":"Sejal Ghate, Alberto Santamaría-Pang, I. Tarapov, H. Sair, Craig K. Jones","doi":"10.48550/arXiv.2209.08200","DOIUrl":"https://doi.org/10.48550/arXiv.2209.08200","url":null,"abstract":"Resting state fMRI is an imaging modality which reveals brain activity localization through signal changes, in what is known as Resting State Networks (RSNs). This technique is gaining popularity in neurosurgical pre-planning to visualize the functional regions and assess regional activity. Labeling of rs-fMRI networks require subject-matter expertise and is time consuming, creating a need for an automated classification algorithm. While the impact of AI in medical diagnosis has shown great progress; deploying and maintaining these in a clinical setting is an unmet need. We propose an end-to-end reproducible pipeline which incorporates image processing of rs-fMRI in a cloud-based workflow while using deep learning to automate the classification of RSNs. We have architected a reproducible Azure Machine Learning cloud-based medical imaging concept pipeline for fMRI analysis integrating the popular FMRIB Software Library (FSL) toolkit. To demonstrate a clinical application using a large dataset, we compare three neural network architectures for classification of deeper RSNs derived from processed rs-fMRI. The three algorithms are: an MLP, a 2D projection-based CNN, and a fully 3D CNN classification networks. Each of the net-works was trained on the rs-fMRI back-projected independent components giving>98% accuracy for each classification method.","PeriodicalId":91444,"journal":{"name":"Advances in visual computing : ... international symposium, ISVC ... : proceedings. International Symposium on Visual Computing","volume":"84 1","pages":"275-283"},"PeriodicalIF":0.0,"publicationDate":"2022-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83846951","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Advances in visual computing : ... international symposium, ISVC ... : proceedings. International Symposium on Visual Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1