首页 > 最新文献

arXiv - EE - Image and Video Processing最新文献

英文 中文
NT-ViT: Neural Transcoding Vision Transformers for EEG-to-fMRI Synthesis NT-ViT:用于脑电图-fMRI 合成的神经转码视觉变换器
Pub Date : 2024-09-18 DOI: arxiv-2409.11836
Romeo Lanzino, Federico Fontana, Luigi Cinque, Francesco Scarcello, Atsuto Maki
This paper introduces the Neural Transcoding Vision Transformer (modelname),a generative model designed to estimate high-resolution functional MagneticResonance Imaging (fMRI) samples from simultaneous Electroencephalography (EEG)data. A key feature of modelname is its Domain Matching (DM) sub-module whicheffectively aligns the latent EEG representations with those of fMRI volumes,enhancing the model's accuracy and reliability. Unlike previous methods thattend to struggle with fidelity and reproducibility of images, modelnameaddresses these challenges by ensuring methodological integrity andhigher-quality reconstructions which we showcase through extensive evaluationon two benchmark datasets; modelname outperforms the current state-of-the-artby a significant margin in both cases, e.g. achieving a $10times$ reduction inRMSE and a $3.14times$ increase in SSIM on the Oddball dataset. An ablationstudy also provides insights into the contribution of each component to themodel's overall effectiveness. This development is critical in offering a newapproach to lessen the time and financial constraints typically linked withhigh-resolution brain imaging, thereby aiding in the swift and precisediagnosis of neurological disorders. Although it is not a replacement foractual fMRI but rather a step towards making such imaging more accessible, webelieve that it represents a pivotal advancement in clinical practice andneuroscience research. Code is available aturl{https://github.com/rom42pla/ntvit}.
本文介绍了神经转码视觉转换器(Neural Transcoding Vision Transformer),这是一个生成模型,旨在从同步脑电图(EEG)数据估算高分辨率功能磁共振成像(fMRI)样本。该模型的一个关键特征是其领域匹配(DM)子模块,它能有效地将潜在的脑电图表征与 fMRI 容量的表征相匹配,从而提高模型的准确性和可靠性。与以往在图像的保真度和可重复性方面存在困难的方法不同,modelname通过确保方法的完整性和更高质量的重构来应对这些挑战,我们通过在两个基准数据集上进行广泛评估来展示这些挑战;modelname在两种情况下都以显著的优势超过了当前最先进的方法,例如,在Oddball数据集上,RMSE降低了10美元/倍,SSIM增加了3.14美元/倍。一项消融研究还让我们深入了解了每个组件对模型整体有效性的贡献。这一发展至关重要,它提供了一种新方法来减少通常与高分辨率脑成像相关的时间和经济限制,从而有助于神经系统疾病的快速精确诊断。虽然它不能取代真正的 fMRI,但它是使这种成像技术更加普及的一步,我们相信它代表了临床实践和神经科学研究的一个关键进步。代码请访问:url{https://github.com/rom42pla/ntvit}。
{"title":"NT-ViT: Neural Transcoding Vision Transformers for EEG-to-fMRI Synthesis","authors":"Romeo Lanzino, Federico Fontana, Luigi Cinque, Francesco Scarcello, Atsuto Maki","doi":"arxiv-2409.11836","DOIUrl":"https://doi.org/arxiv-2409.11836","url":null,"abstract":"This paper introduces the Neural Transcoding Vision Transformer (modelname),\u0000a generative model designed to estimate high-resolution functional Magnetic\u0000Resonance Imaging (fMRI) samples from simultaneous Electroencephalography (EEG)\u0000data. A key feature of modelname is its Domain Matching (DM) sub-module which\u0000effectively aligns the latent EEG representations with those of fMRI volumes,\u0000enhancing the model's accuracy and reliability. Unlike previous methods that\u0000tend to struggle with fidelity and reproducibility of images, modelname\u0000addresses these challenges by ensuring methodological integrity and\u0000higher-quality reconstructions which we showcase through extensive evaluation\u0000on two benchmark datasets; modelname outperforms the current state-of-the-art\u0000by a significant margin in both cases, e.g. achieving a $10times$ reduction in\u0000RMSE and a $3.14times$ increase in SSIM on the Oddball dataset. An ablation\u0000study also provides insights into the contribution of each component to the\u0000model's overall effectiveness. This development is critical in offering a new\u0000approach to lessen the time and financial constraints typically linked with\u0000high-resolution brain imaging, thereby aiding in the swift and precise\u0000diagnosis of neurological disorders. Although it is not a replacement for\u0000actual fMRI but rather a step towards making such imaging more accessible, we\u0000believe that it represents a pivotal advancement in clinical practice and\u0000neuroscience research. Code is available at\u0000url{https://github.com/rom42pla/ntvit}.","PeriodicalId":501289,"journal":{"name":"arXiv - EE - Image and Video Processing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142269817","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
World of Forms: Deformable Geometric Templates for One-Shot Surface Meshing in Coronary CT Angiography 形态世界:用于冠状动脉 CT 血管造影中一次性表面网格化的可变形几何模板
Pub Date : 2024-09-18 DOI: arxiv-2409.11837
Rudolf L. M. van Herten, Ioannis Lagogiannis, Jelmer M. Wolterink, Steffen Bruns, Eva R. Meulendijks, Damini Dey, Joris R. de Groot, José P. Henriques, R. Nils Planken, Simone Saitta, Ivana Išgum
Deep learning-based medical image segmentation and surface mesh generationtypically involve a sequential pipeline from image to segmentation to meshes,often requiring large training datasets while making limited use of priorgeometric knowledge. This may lead to topological inconsistencies andsuboptimal performance in low-data regimes. To address these challenges, wepropose a data-efficient deep learning method for direct 3D anatomical objectsurface meshing using geometric priors. Our approach employs a multi-resolutiongraph neural network that operates on a prior geometric template which isdeformed to fit object boundaries of interest. We show how different templatesmay be used for the different surface meshing targets, and introduce a novelmasked autoencoder pretraining strategy for 3D spherical data. The proposedmethod outperforms nnUNet in a one-shot setting for segmentation of thepericardium, left ventricle (LV) cavity and the LV myocardium. Similarly, themethod outperforms other lumen segmentation operating on multi-planarreformatted images. Results further indicate that mesh quality is on par withor improves upon marching cubes post-processing of voxel mask predictions,while remaining flexible in the choice of mesh triangulation prior, thus pavingthe way for more accurate and topologically consistent 3D medical objectsurface meshing.
基于深度学习的医学图像分割和表面网格生成通常涉及从图像到分割再到网格的顺序流水线,往往需要大量的训练数据集,同时对先验几何知识的利用有限。这可能会导致拓扑不一致,以及在低数据量情况下性能不佳。为了应对这些挑战,我们提出了一种数据高效的深度学习方法,利用几何先验知识直接对三维解剖物体表面进行网格划分。我们的方法采用了多分辨率图神经网络,该网络在先验几何模板上运行,模板经过变形以适合感兴趣的物体边界。我们展示了不同的模板如何用于不同的表面网格划分目标,并针对三维球形数据引入了一种新颖的掩码自动编码器预训练策略。在对心包、左心室(LV)腔和左心室心肌进行一次性分割时,所提出的方法优于 nnUNet。同样,该方法在多平面格式图像上的表现也优于其他管腔分割方法。结果进一步表明,网格质量与行进立方体后处理体素掩模预测不相上下,甚至更胜一筹,同时还能灵活选择网格三角先验,从而为更精确、拓扑更一致的三维医学物体表面网格划分铺平了道路。
{"title":"World of Forms: Deformable Geometric Templates for One-Shot Surface Meshing in Coronary CT Angiography","authors":"Rudolf L. M. van Herten, Ioannis Lagogiannis, Jelmer M. Wolterink, Steffen Bruns, Eva R. Meulendijks, Damini Dey, Joris R. de Groot, José P. Henriques, R. Nils Planken, Simone Saitta, Ivana Išgum","doi":"arxiv-2409.11837","DOIUrl":"https://doi.org/arxiv-2409.11837","url":null,"abstract":"Deep learning-based medical image segmentation and surface mesh generation\u0000typically involve a sequential pipeline from image to segmentation to meshes,\u0000often requiring large training datasets while making limited use of prior\u0000geometric knowledge. This may lead to topological inconsistencies and\u0000suboptimal performance in low-data regimes. To address these challenges, we\u0000propose a data-efficient deep learning method for direct 3D anatomical object\u0000surface meshing using geometric priors. Our approach employs a multi-resolution\u0000graph neural network that operates on a prior geometric template which is\u0000deformed to fit object boundaries of interest. We show how different templates\u0000may be used for the different surface meshing targets, and introduce a novel\u0000masked autoencoder pretraining strategy for 3D spherical data. The proposed\u0000method outperforms nnUNet in a one-shot setting for segmentation of the\u0000pericardium, left ventricle (LV) cavity and the LV myocardium. Similarly, the\u0000method outperforms other lumen segmentation operating on multi-planar\u0000reformatted images. Results further indicate that mesh quality is on par with\u0000or improves upon marching cubes post-processing of voxel mask predictions,\u0000while remaining flexible in the choice of mesh triangulation prior, thus paving\u0000the way for more accurate and topologically consistent 3D medical object\u0000surface meshing.","PeriodicalId":501289,"journal":{"name":"arXiv - EE - Image and Video Processing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142263131","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Autopet III challenge: Incorporating anatomical knowledge into nnUNet for lesion segmentation in PET/CT Autopet III 挑战赛:将解剖学知识纳入 nnUNet,在 PET/CT 中进行病灶分割
Pub Date : 2024-09-18 DOI: arxiv-2409.12155
Hamza Kalisch, Fabian Hörst, Ken Herrmann, Jens Kleesiek, Constantin Seibold
Lesion segmentation in PET/CT imaging is essential for precise tumorcharacterization, which supports personalized treatment planning and enhancesdiagnostic precision in oncology. However, accurate manual segmentation oflesions is time-consuming and prone to inter-observer variability. Given therising demand and clinical use of PET/CT, automated segmentation methods,particularly deep-learning-based approaches, have become increasingly morerelevant. The autoPET III Challenge focuses on advancing automated segmentationof tumor lesions in PET/CT images in a multitracer multicenter setting,addressing the clinical need for quantitative, robust, and generalizablesolutions. Building on previous challenges, the third iteration of the autoPETchallenge introduces a more diverse dataset featuring two different tracers(FDG and PSMA) from two clinical centers. To this extent, we developed aclassifier that identifies the tracer of the given PET/CT based on the MaximumIntensity Projection of the PET scan. We trained two individualnnUNet-ensembles for each tracer where anatomical labels are included as amulti-label task to enhance the model's performance. Our final submissionachieves cross-validation Dice scores of 76.90% and 61.33% for the publiclyavailable FDG and PSMA datasets, respectively. The code is available athttps://github.com/hakal104/autoPETIII/ .
PET/CT 成像中的病灶分割对精确的肿瘤定性至关重要,可支持个性化治疗计划并提高肿瘤诊断的精确性。然而,精确的手动病灶分割既费时又容易造成观察者之间的差异。鉴于 PET/CT 的需求和临床应用日益增长,自动分割方法,尤其是基于深度学习的方法,变得越来越重要。autoPET III 挑战赛的重点是在多示踪剂多中心环境中推进 PET/CT 图像中肿瘤病灶的自动分割,满足临床对定量、稳健和通用解决方案的需求。在前几届挑战赛的基础上,第三届 autoPETchallenge 引入了更多样化的数据集,包括来自两个临床中心的两种不同示踪剂(FDG 和 PSMA)。为此,我们开发了一种分类器,可根据 PET 扫描的最大强度投影来识别给定 PET/CT 的示踪剂。我们为每种示踪剂训练了两个独立的 nnUNet 集合,其中解剖学标签被列为多标签任务,以提高模型的性能。对于公开的 FDG 和 PSMA 数据集,我们最终提交的交叉验证 Dice 分数分别为 76.90% 和 61.33%。代码可在https://github.com/hakal104/autoPETIII/。
{"title":"Autopet III challenge: Incorporating anatomical knowledge into nnUNet for lesion segmentation in PET/CT","authors":"Hamza Kalisch, Fabian Hörst, Ken Herrmann, Jens Kleesiek, Constantin Seibold","doi":"arxiv-2409.12155","DOIUrl":"https://doi.org/arxiv-2409.12155","url":null,"abstract":"Lesion segmentation in PET/CT imaging is essential for precise tumor\u0000characterization, which supports personalized treatment planning and enhances\u0000diagnostic precision in oncology. However, accurate manual segmentation of\u0000lesions is time-consuming and prone to inter-observer variability. Given the\u0000rising demand and clinical use of PET/CT, automated segmentation methods,\u0000particularly deep-learning-based approaches, have become increasingly more\u0000relevant. The autoPET III Challenge focuses on advancing automated segmentation\u0000of tumor lesions in PET/CT images in a multitracer multicenter setting,\u0000addressing the clinical need for quantitative, robust, and generalizable\u0000solutions. Building on previous challenges, the third iteration of the autoPET\u0000challenge introduces a more diverse dataset featuring two different tracers\u0000(FDG and PSMA) from two clinical centers. To this extent, we developed a\u0000classifier that identifies the tracer of the given PET/CT based on the Maximum\u0000Intensity Projection of the PET scan. We trained two individual\u0000nnUNet-ensembles for each tracer where anatomical labels are included as a\u0000multi-label task to enhance the model's performance. Our final submission\u0000achieves cross-validation Dice scores of 76.90% and 61.33% for the publicly\u0000available FDG and PSMA datasets, respectively. The code is available at\u0000https://github.com/hakal104/autoPETIII/ .","PeriodicalId":501289,"journal":{"name":"arXiv - EE - Image and Video Processing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142263127","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Few-Shot Learning Approach on Tuberculosis Classification Based on Chest X-Ray Images 基于胸部 X 射线图像的肺结核分类 "少量学习 "方法
Pub Date : 2024-09-18 DOI: arxiv-2409.11644
A. A. G. Yogi Pramana, Faiz Ihza Permana, Muhammad Fazil Maulana, Dzikri Rahadian Fudholi
Tuberculosis (TB) is caused by the bacterium Mycobacterium tuberculosis,primarily affecting the lungs. Early detection is crucial for improvingtreatment effectiveness and reducing transmission risk. Artificial intelligence(AI), particularly through image classification of chest X-rays, can assist inTB detection. However, class imbalance in TB chest X-ray datasets presents achallenge for accurate classification. In this paper, we propose a few-shotlearning (FSL) approach using the Prototypical Network algorithm to addressthis issue. We compare the performance of ResNet-18, ResNet-50, and VGG16 infeature extraction from the TBX11K Chest X-ray dataset. Experimental resultsdemonstrate classification accuracies of 98.93% for ResNet-18, 98.60% forResNet-50, and 33.33% for VGG16. These findings indicate that the proposedmethod outperforms others in mitigating data imbalance, which is particularlybeneficial for disease classification applications.
结核病(TB)是由结核分枝杆菌引起的,主要侵犯肺部。早期发现对于提高治疗效果和降低传播风险至关重要。人工智能(AI),尤其是通过对胸部 X 光片进行图像分类,可以帮助检测结核病。然而,结核病胸部 X 光片数据集中的类不平衡给准确分类带来了挑战。在本文中,我们提出了一种使用原型网络算法的 "少量清除"(FSL)方法来解决这一问题。我们比较了 ResNet-18、ResNet-50 和 VGG16 从 TBX11K 胸部 X 光数据集中提取特征的性能。实验结果表明,ResNet-18 的分类准确率为 98.93%,ResNet-50 为 98.60%,VGG16 为 33.33%。这些结果表明,所提出的方法在缓解数据不平衡方面优于其他方法,这对疾病分类应用尤其有益。
{"title":"Few-Shot Learning Approach on Tuberculosis Classification Based on Chest X-Ray Images","authors":"A. A. G. Yogi Pramana, Faiz Ihza Permana, Muhammad Fazil Maulana, Dzikri Rahadian Fudholi","doi":"arxiv-2409.11644","DOIUrl":"https://doi.org/arxiv-2409.11644","url":null,"abstract":"Tuberculosis (TB) is caused by the bacterium Mycobacterium tuberculosis,\u0000primarily affecting the lungs. Early detection is crucial for improving\u0000treatment effectiveness and reducing transmission risk. Artificial intelligence\u0000(AI), particularly through image classification of chest X-rays, can assist in\u0000TB detection. However, class imbalance in TB chest X-ray datasets presents a\u0000challenge for accurate classification. In this paper, we propose a few-shot\u0000learning (FSL) approach using the Prototypical Network algorithm to address\u0000this issue. We compare the performance of ResNet-18, ResNet-50, and VGG16 in\u0000feature extraction from the TBX11K Chest X-ray dataset. Experimental results\u0000demonstrate classification accuracies of 98.93% for ResNet-18, 98.60% for\u0000ResNet-50, and 33.33% for VGG16. These findings indicate that the proposed\u0000method outperforms others in mitigating data imbalance, which is particularly\u0000beneficial for disease classification applications.","PeriodicalId":501289,"journal":{"name":"arXiv - EE - Image and Video Processing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142269816","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ABHINAW: A method for Automatic Evaluation of Typography within AI-Generated Images ABHINAW:自动评估人工智能生成图像中排版的方法
Pub Date : 2024-09-18 DOI: arxiv-2409.11874
Abhinaw Jagtap, Nachiket Tapas, R. G. Brajesh
In the fast-evolving field of Generative AI, platforms like MidJourney,DALL-E, and Stable Diffusion have transformed Text-to-Image (T2I) Generation.However, despite their impressive ability to create high-quality images, theyoften struggle to generate accurate text within these images. Theoretically, ifwe could achieve accurate text generation in AI images in a ``zero-shot''manner, it would not only make AI-generated images more meaningful but alsodemocratize the graphic design industry. The first step towards this goal is tocreate a robust scoring matrix for evaluating text accuracy in AI-generatedimages. Although there are existing bench-marking methods like CLIP SCORE andT2I-CompBench++, there's still a gap in systematically evaluating text andtypography in AI-generated images, especially with diffusion-based methods. Inthis paper, we introduce a novel evaluation matrix designed explicitly forquantifying the performance of text and typography generation withinAI-generated images. We have used letter by letter matching strategy to computethe exact matching scores from the reference text to the AI generated text. Ournovel approach to calculate the score takes care of multiple redundancies suchas repetition of words, case sensitivity, mixing of words, irregularincorporation of letters etc. Moreover, we have developed a Novel method namedas brevity adjustment to handle excess text. In addition we have also done aquantitative analysis of frequent errors arise due to frequently used words andless frequently used words. Project page is available at:https://github.com/Abhinaw3906/ABHINAW-MATRIX.
在快速发展的生成式人工智能领域,MidJourney、DALL-E 和 Stable Diffusion 等平台已经改变了文本到图像(T2I)的生成方式。然而,尽管这些平台具有令人印象深刻的创建高质量图像的能力,但它们往往难以在这些图像中生成准确的文本。从理论上讲,如果我们能以 "零误差 "的方式在人工智能图像中实现准确的文本生成,这不仅会使人工智能生成的图像更有意义,而且还会使平面设计行业民主化。实现这一目标的第一步是创建一个强大的评分矩阵,用于评估人工智能生成图像中文字的准确性。虽然已有 CLIP SCORE 和T2I-CompBench++ 等基准标记方法,但在系统评估人工智能生成图像中的文字和排版方面仍存在差距,尤其是基于扩散的方法。在本文中,我们引入了一个新颖的评估矩阵,专门用于量化人工智能生成的图像中文本和排版生成的性能。我们采用逐个字母匹配的策略来计算参考文本与人工智能生成文本的精确匹配分数。我们计算分数的新方法考虑到了多种冗余,如单词重复、大小写敏感性、单词混合、字母的不规则合并等。此外,我们还开发了一种名为 "简短度调整 "的新方法来处理多余的文本。此外,我们还对因常用词和非常用词而产生的常见错误进行了定量分析。项目网页:https://github.com/Abhinaw3906/ABHINAW-MATRIX。
{"title":"ABHINAW: A method for Automatic Evaluation of Typography within AI-Generated Images","authors":"Abhinaw Jagtap, Nachiket Tapas, R. G. Brajesh","doi":"arxiv-2409.11874","DOIUrl":"https://doi.org/arxiv-2409.11874","url":null,"abstract":"In the fast-evolving field of Generative AI, platforms like MidJourney,\u0000DALL-E, and Stable Diffusion have transformed Text-to-Image (T2I) Generation.\u0000However, despite their impressive ability to create high-quality images, they\u0000often struggle to generate accurate text within these images. Theoretically, if\u0000we could achieve accurate text generation in AI images in a ``zero-shot''\u0000manner, it would not only make AI-generated images more meaningful but also\u0000democratize the graphic design industry. The first step towards this goal is to\u0000create a robust scoring matrix for evaluating text accuracy in AI-generated\u0000images. Although there are existing bench-marking methods like CLIP SCORE and\u0000T2I-CompBench++, there's still a gap in systematically evaluating text and\u0000typography in AI-generated images, especially with diffusion-based methods. In\u0000this paper, we introduce a novel evaluation matrix designed explicitly for\u0000quantifying the performance of text and typography generation within\u0000AI-generated images. We have used letter by letter matching strategy to compute\u0000the exact matching scores from the reference text to the AI generated text. Our\u0000novel approach to calculate the score takes care of multiple redundancies such\u0000as repetition of words, case sensitivity, mixing of words, irregular\u0000incorporation of letters etc. Moreover, we have developed a Novel method named\u0000as brevity adjustment to handle excess text. In addition we have also done a\u0000quantitative analysis of frequent errors arise due to frequently used words and\u0000less frequently used words. Project page is available at:\u0000https://github.com/Abhinaw3906/ABHINAW-MATRIX.","PeriodicalId":501289,"journal":{"name":"arXiv - EE - Image and Video Processing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142269823","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-Sensor Deep Learning for Glacier Mapping 冰川测绘的多传感器深度学习
Pub Date : 2024-09-18 DOI: arxiv-2409.12034
Codruţ-Andrei Diaconu, Konrad Heidler, Jonathan L. Bamber, Harry Zekollari
The more than 200,000 glaciers outside the ice sheets play a crucial role inour society by influencing sea-level rise, water resource management, naturalhazards, biodiversity, and tourism. However, only a fraction of these glaciersbenefit from consistent and detailed in-situ observations that allow forassessing their status and changes over time. This limitation can, in part, beovercome by relying on satellite-based Earth Observation techniques.Satellite-based glacier mapping applications have historically mainly relied onmanual and semi-automatic detection methods, while recently, a fast and notabletransition to deep learning techniques has started. This chapter reviews how combining multi-sensor remote sensing data and deeplearning allows us to better delineate (i.e. map) glaciers and detect theirtemporal changes. We explain how relying on deep learning multi-sensorframeworks to map glaciers benefits from the extensive availability of regionaland global glacier inventories. We also analyse the rationale behind glaciermapping, the benefits of deep learning methodologies, and the inherentchallenges in integrating multi-sensor earth observation data with deeplearning algorithms. While our review aims to provide a broad overview of glacier mapping efforts,we highlight a few setups where deep learning multi-sensor remote sensingapplications have a considerable potential added value. This includesapplications for debris-covered and rock glaciers that are visually difficultto distinguish from surroundings and for calving glaciers that are in contactwith the ocean. These specific cases are illustrated through a series of visualimageries, highlighting some significant advantages and challenges whendetecting glacier changes, including dealing with seasonal snow cover, changingdebris coverage, and distinguishing glacier fronts from the surrounding seaice.
冰原之外的 20 多万个冰川对海平面上升、水资源管理、自然灾害、生物多样性和旅游业都有影响,在我们的社会中发挥着至关重要的作用。然而,只有一小部分冰川受益于持续、详细的现场观测,从而能够评估其状态和随时间的变化。基于卫星的冰川测绘应用历来主要依赖人工和半自动探测方法,而最近开始向深度学习技术快速而显著地过渡。本章回顾了多传感器遥感数据与深度学习的结合如何让我们更好地划分(即绘制)冰川并探测其时变。我们解释了依靠深度学习多传感器框架绘制冰川地图如何受益于区域和全球冰川清单的广泛可用性。我们还分析了冰川测绘背后的原理、深度学习方法的益处,以及将多传感器地球观测数据与深度学习算法相结合的固有挑战。虽然我们的综述旨在对冰川测绘工作进行广泛概述,但我们强调了深度学习多传感器遥感应用具有相当大潜在附加值的一些设置。这包括应用于从视觉上很难与周围环境区分开来的碎屑覆盖冰川和岩石冰川,以及与海洋接触的塌方冰川。这些具体案例通过一系列可视成像进行了说明,突出了在检测冰川变化时的一些重要优势和挑战,包括处理季节性积雪、不断变化的碎石覆盖以及区分冰川前沿和周围海冰。
{"title":"Multi-Sensor Deep Learning for Glacier Mapping","authors":"Codruţ-Andrei Diaconu, Konrad Heidler, Jonathan L. Bamber, Harry Zekollari","doi":"arxiv-2409.12034","DOIUrl":"https://doi.org/arxiv-2409.12034","url":null,"abstract":"The more than 200,000 glaciers outside the ice sheets play a crucial role in\u0000our society by influencing sea-level rise, water resource management, natural\u0000hazards, biodiversity, and tourism. However, only a fraction of these glaciers\u0000benefit from consistent and detailed in-situ observations that allow for\u0000assessing their status and changes over time. This limitation can, in part, be\u0000overcome by relying on satellite-based Earth Observation techniques.\u0000Satellite-based glacier mapping applications have historically mainly relied on\u0000manual and semi-automatic detection methods, while recently, a fast and notable\u0000transition to deep learning techniques has started. This chapter reviews how combining multi-sensor remote sensing data and deep\u0000learning allows us to better delineate (i.e. map) glaciers and detect their\u0000temporal changes. We explain how relying on deep learning multi-sensor\u0000frameworks to map glaciers benefits from the extensive availability of regional\u0000and global glacier inventories. We also analyse the rationale behind glacier\u0000mapping, the benefits of deep learning methodologies, and the inherent\u0000challenges in integrating multi-sensor earth observation data with deep\u0000learning algorithms. While our review aims to provide a broad overview of glacier mapping efforts,\u0000we highlight a few setups where deep learning multi-sensor remote sensing\u0000applications have a considerable potential added value. This includes\u0000applications for debris-covered and rock glaciers that are visually difficult\u0000to distinguish from surroundings and for calving glaciers that are in contact\u0000with the ocean. These specific cases are illustrated through a series of visual\u0000imageries, highlighting some significant advantages and challenges when\u0000detecting glacier changes, including dealing with seasonal snow cover, changing\u0000debris coverage, and distinguishing glacier fronts from the surrounding sea\u0000ice.","PeriodicalId":501289,"journal":{"name":"arXiv - EE - Image and Video Processing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142269821","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LFIC-DRASC: Deep Light Field Image Compression Using Disentangled Representation and Asymmetrical Strip Convolution LFIC-DRASC:利用分离表示和非对称条带卷积进行深度光场图像压缩
Pub Date : 2024-09-18 DOI: arxiv-2409.11711
Shiyu Feng, Yun Zhang, Linwei Zhu, Sam Kwong
Light-Field (LF) image is emerging 4D data of light rays that is capable ofrealistically presenting spatial and angular information of 3D scene. However,the large data volume of LF images becomes the most challenging issue inreal-time processing, transmission, and storage. In this paper, we propose anend-to-end deep LF Image Compression method Using Disentangled Representationand Asymmetrical Strip Convolution (LFIC-DRASC) to improve coding efficiency.Firstly, we formulate the LF image compression problem as learning adisentangled LF representation network and an image encoding-decoding network.Secondly, we propose two novel feature extractors that leverage the structuralprior of LF data by integrating features across different dimensions.Meanwhile, disentangled LF representation network is proposed to enhance the LFfeature disentangling and decoupling. Thirdly, we propose the LFIC-DRASC for LFimage compression, where two Asymmetrical Strip Convolution (ASC) operators,i.e. horizontal and vertical, are proposed to capture long-range correlation inLF feature space. These two ASC operators can be combined with the squareconvolution to further decouple LF features, which enhances the model abilityin representing intricate spatial relationships. Experimental resultsdemonstrate that the proposed LFIC-DRASC achieves an average of 20.5% bit ratereductions comparing with the state-of-the-art methods.
光场(LF)图像是新兴的光射线四维数据,能够真实呈现三维场景的空间和角度信息。然而,光场图像数据量大,成为实时处理、传输和存储中最具挑战性的问题。本文提出了一种端到端的深度低频图像压缩方法--使用非平行表示和非对称条带卷积(LFIC-DRASC)来提高编码效率。其次,我们提出了两个新颖的特征提取器,通过整合不同维度的特征来利用低频数据的结构先验性。第三,我们提出了用于低频图像压缩的 LFIC-DRASC,其中提出了两个非对称带卷积(ASC)算子,即水平和垂直算子,以捕捉低频特征空间中的长程相关性。这两个 ASC 算子可与平方卷积相结合,进一步解耦低频特征,从而增强了模型表现复杂空间关系的能力。实验结果表明,与最先进的方法相比,所提出的 LFIC-DRASC 平均降低了 20.5% 的比特率。
{"title":"LFIC-DRASC: Deep Light Field Image Compression Using Disentangled Representation and Asymmetrical Strip Convolution","authors":"Shiyu Feng, Yun Zhang, Linwei Zhu, Sam Kwong","doi":"arxiv-2409.11711","DOIUrl":"https://doi.org/arxiv-2409.11711","url":null,"abstract":"Light-Field (LF) image is emerging 4D data of light rays that is capable of\u0000realistically presenting spatial and angular information of 3D scene. However,\u0000the large data volume of LF images becomes the most challenging issue in\u0000real-time processing, transmission, and storage. In this paper, we propose an\u0000end-to-end deep LF Image Compression method Using Disentangled Representation\u0000and Asymmetrical Strip Convolution (LFIC-DRASC) to improve coding efficiency.\u0000Firstly, we formulate the LF image compression problem as learning a\u0000disentangled LF representation network and an image encoding-decoding network.\u0000Secondly, we propose two novel feature extractors that leverage the structural\u0000prior of LF data by integrating features across different dimensions.\u0000Meanwhile, disentangled LF representation network is proposed to enhance the LF\u0000feature disentangling and decoupling. Thirdly, we propose the LFIC-DRASC for LF\u0000image compression, where two Asymmetrical Strip Convolution (ASC) operators,\u0000i.e. horizontal and vertical, are proposed to capture long-range correlation in\u0000LF feature space. These two ASC operators can be combined with the square\u0000convolution to further decouple LF features, which enhances the model ability\u0000in representing intricate spatial relationships. Experimental results\u0000demonstrate that the proposed LFIC-DRASC achieves an average of 20.5% bit rate\u0000reductions comparing with the state-of-the-art methods.","PeriodicalId":501289,"journal":{"name":"arXiv - EE - Image and Video Processing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142263132","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
multiPI-TransBTS: A Multi-Path Learning Framework for Brain Tumor Image Segmentation Based on Multi-Physical Information multiPI-TransBTS:基于多物理信息的脑肿瘤图像分割多路径学习框架
Pub Date : 2024-09-18 DOI: arxiv-2409.12167
Hongjun Zhu, Jiaohang Huang, Kuo Chen, Xuehui Ying, Ying Qian
Brain Tumor Segmentation (BraTS) plays a critical role in clinical diagnosis,treatment planning, and monitoring the progression of brain tumors. However,due to the variability in tumor appearance, size, and intensity acrossdifferent MRI modalities, automated segmentation remains a challenging task. Inthis study, we propose a novel Transformer-based framework, multiPI-TransBTS,which integrates multi-physical information to enhance segmentation accuracy.The model leverages spatial information, semantic information, and multi-modalimaging data, addressing the inherent heterogeneity in brain tumorcharacteristics. The multiPI-TransBTS framework consists of an encoder, anAdaptive Feature Fusion (AFF) module, and a multi-source, multi-scale featuredecoder. The encoder incorporates a multi-branch architecture to separatelyextract modality-specific features from different MRI sequences. The AFF modulefuses information from multiple sources using channel-wise and element-wiseattention, ensuring effective feature recalibration. The decoder combines bothcommon and task-specific features through a Task-Specific Feature Introduction(TSFI) strategy, producing accurate segmentation outputs for Whole Tumor (WT),Tumor Core (TC), and Enhancing Tumor (ET) regions. Comprehensive evaluations onthe BraTS2019 and BraTS2020 datasets demonstrate the superiority ofmultiPI-TransBTS over the state-of-the-art methods. The model consistentlyachieves better Dice coefficients, Hausdorff distances, and Sensitivity scores,highlighting its effectiveness in addressing the BraTS challenges. Our resultsalso indicate the need for further exploration of the balance between precisionand recall in the ET segmentation task. The proposed framework represents asignificant advancement in BraTS, with potential implications for improvingclinical outcomes for brain tumor patients.
脑肿瘤分割(Brain Tumor Segmentation,BraTS)在临床诊断、治疗计划和监测脑肿瘤进展方面发挥着至关重要的作用。然而,由于不同磁共振成像模式下肿瘤的外观、大小和强度存在差异,自动分割仍是一项具有挑战性的任务。该模型利用空间信息、语义信息和多模态成像数据,解决了脑肿瘤固有的异质性特征。multiPI-TransBTS 框架由编码器、自适应特征融合(AFF)模块和多源多尺度特征编码器组成。编码器采用多分支架构,分别从不同的磁共振成像序列中提取特定模态的特征。AFF 模块利用信道和元素注意力融合来自多个来源的信息,确保有效的特征重新校准。解码器通过任务特异性特征引入(TSFI)策略将通用特征和任务特异性特征结合起来,为整个肿瘤(WT)、肿瘤核心(TC)和增强肿瘤(ET)区域提供精确的分割输出。在 BraTS2019 和 BraTS2020 数据集上进行的综合评估表明,multiPI-TransBTS 优于最先进的方法。该模型的 Dice 系数、Hausdorff 距离和灵敏度得分一直较高,突出表明了它在应对 BraTS 挑战方面的有效性。我们的结果还表明,有必要进一步探索在 ET 分割任务中精度和召回率之间的平衡。所提出的框架代表了 BraTS 的重大进步,对改善脑肿瘤患者的临床疗效具有潜在意义。
{"title":"multiPI-TransBTS: A Multi-Path Learning Framework for Brain Tumor Image Segmentation Based on Multi-Physical Information","authors":"Hongjun Zhu, Jiaohang Huang, Kuo Chen, Xuehui Ying, Ying Qian","doi":"arxiv-2409.12167","DOIUrl":"https://doi.org/arxiv-2409.12167","url":null,"abstract":"Brain Tumor Segmentation (BraTS) plays a critical role in clinical diagnosis,\u0000treatment planning, and monitoring the progression of brain tumors. However,\u0000due to the variability in tumor appearance, size, and intensity across\u0000different MRI modalities, automated segmentation remains a challenging task. In\u0000this study, we propose a novel Transformer-based framework, multiPI-TransBTS,\u0000which integrates multi-physical information to enhance segmentation accuracy.\u0000The model leverages spatial information, semantic information, and multi-modal\u0000imaging data, addressing the inherent heterogeneity in brain tumor\u0000characteristics. The multiPI-TransBTS framework consists of an encoder, an\u0000Adaptive Feature Fusion (AFF) module, and a multi-source, multi-scale feature\u0000decoder. The encoder incorporates a multi-branch architecture to separately\u0000extract modality-specific features from different MRI sequences. The AFF module\u0000fuses information from multiple sources using channel-wise and element-wise\u0000attention, ensuring effective feature recalibration. The decoder combines both\u0000common and task-specific features through a Task-Specific Feature Introduction\u0000(TSFI) strategy, producing accurate segmentation outputs for Whole Tumor (WT),\u0000Tumor Core (TC), and Enhancing Tumor (ET) regions. Comprehensive evaluations on\u0000the BraTS2019 and BraTS2020 datasets demonstrate the superiority of\u0000multiPI-TransBTS over the state-of-the-art methods. The model consistently\u0000achieves better Dice coefficients, Hausdorff distances, and Sensitivity scores,\u0000highlighting its effectiveness in addressing the BraTS challenges. Our results\u0000also indicate the need for further exploration of the balance between precision\u0000and recall in the ET segmentation task. The proposed framework represents a\u0000significant advancement in BraTS, with potential implications for improving\u0000clinical outcomes for brain tumor patients.","PeriodicalId":501289,"journal":{"name":"arXiv - EE - Image and Video Processing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142263126","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Denoising diffusion models for high-resolution microscopy image restoration 用于高分辨率显微图像复原的去噪扩散模型
Pub Date : 2024-09-18 DOI: arxiv-2409.12078
Pamela Osuna-Vargas, Maren H. Wehrheim, Lucas Zinz, Johanna Rahm, Ashwin Balakrishnan, Alexandra Kaminer, Mike Heilemann, Matthias Kaschube
Advances in microscopy imaging enable researchers to visualize structures atthe nanoscale level thereby unraveling intricate details of biologicalorganization. However, challenges such as image noise, photobleaching offluorophores, and low tolerability of biological samples to high light dosesremain, restricting temporal resolutions and experiment durations. Reducedlaser doses enable longer measurements at the cost of lower resolution andincreased noise, which hinders accurate downstream analyses. Here we train adenoising diffusion probabilistic model (DDPM) to predict high-resolutionimages by conditioning the model on low-resolution information. Additionally,the probabilistic aspect of the DDPM allows for repeated generation of imagesthat tend to further increase the signal-to-noise ratio. We show that our modelachieves a performance that is better or similar to the previouslybest-performing methods, across four highly diverse datasets. Importantly,while any of the previous methods show competitive performance for some, butnot all datasets, our method consistently achieves high performance across allfour data sets, suggesting high generalizability.
显微成像技术的进步使研究人员能够在纳米级水平上观察结构,从而揭示生物组织的复杂细节。然而,图像噪声、荧光团的光漂白、生物样本对高光剂量的耐受性低等挑战依然存在,限制了时间分辨率和实验持续时间。降低激光剂量可以延长测量时间,但代价是降低分辨率和增加噪声,从而阻碍了下游分析的准确性。在这里,我们通过在低分辨率信息的基础上训练腺扩散概率模型(DDPM)来预测高分辨率图像。此外,DDPM 的概率方面允许重复生成图像,从而进一步提高信噪比。我们的研究表明,在四个高度多样化的数据集上,我们的模型取得了优于或类似于之前表现最好的方法的性能。重要的是,虽然之前的任何方法都能在某些数据集(而非所有数据集)上显示出具有竞争力的性能,但我们的方法却能在所有四个数据集上持续获得高性能,这表明我们的方法具有很高的通用性。
{"title":"Denoising diffusion models for high-resolution microscopy image restoration","authors":"Pamela Osuna-Vargas, Maren H. Wehrheim, Lucas Zinz, Johanna Rahm, Ashwin Balakrishnan, Alexandra Kaminer, Mike Heilemann, Matthias Kaschube","doi":"arxiv-2409.12078","DOIUrl":"https://doi.org/arxiv-2409.12078","url":null,"abstract":"Advances in microscopy imaging enable researchers to visualize structures at\u0000the nanoscale level thereby unraveling intricate details of biological\u0000organization. However, challenges such as image noise, photobleaching of\u0000fluorophores, and low tolerability of biological samples to high light doses\u0000remain, restricting temporal resolutions and experiment durations. Reduced\u0000laser doses enable longer measurements at the cost of lower resolution and\u0000increased noise, which hinders accurate downstream analyses. Here we train a\u0000denoising diffusion probabilistic model (DDPM) to predict high-resolution\u0000images by conditioning the model on low-resolution information. Additionally,\u0000the probabilistic aspect of the DDPM allows for repeated generation of images\u0000that tend to further increase the signal-to-noise ratio. We show that our model\u0000achieves a performance that is better or similar to the previously\u0000best-performing methods, across four highly diverse datasets. Importantly,\u0000while any of the previous methods show competitive performance for some, but\u0000not all datasets, our method consistently achieves high performance across all\u0000four data sets, suggesting high generalizability.","PeriodicalId":501289,"journal":{"name":"arXiv - EE - Image and Video Processing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142263128","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Computational Imaging for Long-Term Prediction of Solar Irradiance 用于长期预测太阳辐照度的计算成像技术
Pub Date : 2024-09-18 DOI: arxiv-2409.12016
Leron Julian, Haejoon Lee, Soummya Kar, Aswin C. Sankaranarayanan
The occlusion of the sun by clouds is one of the primary sources ofuncertainties in solar power generation, and is a factor that affects thewide-spread use of solar power as a primary energy source. Real-timeforecasting of cloud movement and, as a result, solar irradiance is necessaryto schedule and allocate energy across grid-connected photovoltaic systems.Previous works monitored cloud movement using wide-angle field of view imageryof the sky. However, such images have poor resolution for clouds that appearnear the horizon, which reduces their effectiveness for long term prediction ofsolar occlusion. Specifically, to be able to predict occlusion of the sun overlong time periods, clouds that are near the horizon need to be detected, andtheir velocities estimated precisely. To enable such a system, we design anddeploy a catadioptric system that delivers wide-angle imagery with uniformspatial resolution of the sky over its field of view. To enable prediction overa longer time horizon, we design an algorithm that uses carefully selectedspatio-temporal slices of the imagery using estimated wind direction andvelocity as inputs. Using ray-tracing simulations as well as a real testbeddeployed outdoors, we show that the system is capable of predicting solarocclusion as well as irradiance for tens of minutes in the future, which is anorder of magnitude improvement over prior work.
云层遮挡太阳是太阳能发电不确定性的主要来源之一,也是影响将太阳能作为主要能源广泛使用的一个因素。对云层移动以及由此产生的太阳辐照度进行实时预测,对并网光伏系统的能源调度和分配十分必要。然而,这些图像对出现在地平线附近的云的分辨率较低,这降低了它们对长期预测太阳遮挡的有效性。具体来说,为了能够预测长时间的太阳遮挡,需要检测地平线附近的云层,并精确估计它们的速度。为了实现这样一个系统,我们设计并部署了一个视场内天空空间分辨率一致的广角成像系统。为了能够对更长的时间范围进行预测,我们设计了一种算法,该算法以估计的风向和风速为输入,使用精心选择的图像时空切片。通过射线追踪模拟以及在室外部署的真实测试平台,我们证明该系统能够预测未来数十分钟内的太阳闭塞度和辐照度,这比之前的工作有了数量级的提升。
{"title":"Computational Imaging for Long-Term Prediction of Solar Irradiance","authors":"Leron Julian, Haejoon Lee, Soummya Kar, Aswin C. Sankaranarayanan","doi":"arxiv-2409.12016","DOIUrl":"https://doi.org/arxiv-2409.12016","url":null,"abstract":"The occlusion of the sun by clouds is one of the primary sources of\u0000uncertainties in solar power generation, and is a factor that affects the\u0000wide-spread use of solar power as a primary energy source. Real-time\u0000forecasting of cloud movement and, as a result, solar irradiance is necessary\u0000to schedule and allocate energy across grid-connected photovoltaic systems.\u0000Previous works monitored cloud movement using wide-angle field of view imagery\u0000of the sky. However, such images have poor resolution for clouds that appear\u0000near the horizon, which reduces their effectiveness for long term prediction of\u0000solar occlusion. Specifically, to be able to predict occlusion of the sun over\u0000long time periods, clouds that are near the horizon need to be detected, and\u0000their velocities estimated precisely. To enable such a system, we design and\u0000deploy a catadioptric system that delivers wide-angle imagery with uniform\u0000spatial resolution of the sky over its field of view. To enable prediction over\u0000a longer time horizon, we design an algorithm that uses carefully selected\u0000spatio-temporal slices of the imagery using estimated wind direction and\u0000velocity as inputs. Using ray-tracing simulations as well as a real testbed\u0000deployed outdoors, we show that the system is capable of predicting solar\u0000occlusion as well as irradiance for tens of minutes in the future, which is an\u0000order of magnitude improvement over prior work.","PeriodicalId":501289,"journal":{"name":"arXiv - EE - Image and Video Processing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142263233","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
arXiv - EE - Image and Video Processing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1