首页 > 最新文献

2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)最新文献

英文 中文
Learning-based compression of visual objects for smart surveillance 基于学习的智能监控视觉对象压缩
Ruben Antonio, S. Faria, Luis M. N. Tavora, A. Navarro, P. Assunção
Advanced video applications in smart environments (e.g., smart cities) bring different challenges associated with increasingly intelligent systems and demanding requirements in emerging fields such as urban surveillance, computer vision in industry, medicine and others. As a consequence, a huge amount of visual data is captured to be analyzed by task-algorithm driven machines. In this context, this paper proposes an efficient learning-based approach to compress relevant visual objects, captured in surveillance contexts and delivered for machine vision processing. An object-based compression scheme is devised, comprising multiple autoencoders, each one optimised to produce an efficient latent representation of a corresponding object class. The performance of the proposed approach is evaluated with two types of visual objects: persons and faces and two task-algorithms: class identification and object recognition, besides traditional image quality metrics like PSNR and VMAF. In comparison with the Versatile Video Coding (VVC) standard, the proposed approach achieves significantly better coding efficiency than the VVC, e.g., up to 46.7% BD-rate reduction. The accuracy of the machine vision tasks is also significantly higher when performed over visual objects compressed with the proposed scheme in comparison with the same tasks performed over the same visual objects compressed with the VVC. These results demonstrate that the learning-based approach proposed in this paper is a more efficient solution for compression of visual objects than standard encoding.
智能环境(例如,智能城市)中的高级视频应用带来了与日益智能的系统和新兴领域(如城市监控,工业计算机视觉,医学等)的苛刻要求相关的不同挑战。因此,大量的视觉数据被捕获,并由任务算法驱动的机器进行分析。在此背景下,本文提出了一种高效的基于学习的方法来压缩相关的视觉对象,这些对象在监视环境中捕获并交付给机器视觉处理。设计了一种基于对象的压缩方案,包括多个自动编码器,每个编码器都经过优化以产生相应对象类的有效潜在表示。除了传统的图像质量指标(如PSNR和VMAF)外,还使用两种类型的视觉对象(人和面孔)以及两种任务算法(类识别和对象识别)来评估该方法的性能。与通用视频编码(VVC)标准相比,该方法的编码效率明显高于VVC标准,可将bd率降低46.7%。与使用VVC压缩的相同视觉对象执行相同的任务相比,使用该方案压缩的视觉对象执行相同的任务时,机器视觉任务的准确性也显着更高。这些结果表明,本文提出的基于学习的方法是一种比标准编码更有效的视觉对象压缩解决方案。
{"title":"Learning-based compression of visual objects for smart surveillance","authors":"Ruben Antonio, S. Faria, Luis M. N. Tavora, A. Navarro, P. Assunção","doi":"10.1109/IPTA54936.2022.9784147","DOIUrl":"https://doi.org/10.1109/IPTA54936.2022.9784147","url":null,"abstract":"Advanced video applications in smart environments (e.g., smart cities) bring different challenges associated with increasingly intelligent systems and demanding requirements in emerging fields such as urban surveillance, computer vision in industry, medicine and others. As a consequence, a huge amount of visual data is captured to be analyzed by task-algorithm driven machines. In this context, this paper proposes an efficient learning-based approach to compress relevant visual objects, captured in surveillance contexts and delivered for machine vision processing. An object-based compression scheme is devised, comprising multiple autoencoders, each one optimised to produce an efficient latent representation of a corresponding object class. The performance of the proposed approach is evaluated with two types of visual objects: persons and faces and two task-algorithms: class identification and object recognition, besides traditional image quality metrics like PSNR and VMAF. In comparison with the Versatile Video Coding (VVC) standard, the proposed approach achieves significantly better coding efficiency than the VVC, e.g., up to 46.7% BD-rate reduction. The accuracy of the machine vision tasks is also significantly higher when performed over visual objects compressed with the proposed scheme in comparison with the same tasks performed over the same visual objects compressed with the VVC. These results demonstrate that the learning-based approach proposed in this paper is a more efficient solution for compression of visual objects than standard encoding.","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126634587","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Spatial Maturity Regression for the Classification of Hematopoietic Cells 用于造血细胞分类的空间成熟度回归
Philipp Gräbel, Julian Thull, M. Crysandt, B. Klinkhammer, P. Boor, T. Brümmendorf, D. Merhof
In contrast to peripheral blood, cells in bone marrow microscopy images are not only characterized by the cell lineage but also a maturity stage within the lineage. As maturation is a continuous process, the differentiation between various stages falls into the category of (ordinal) regression. In this work, we propose Spatial Maturity Regression - a technique that regularizes the learning process to enforce a sensible positioning of maturity stages in the embedding space. To this end, we propose and evaluate several curve models, target definitions and loss function that incorporate this domain knowledge. We show that the classification F-scores improve up to 2.4 percentage points when enforcing regression targets along learnable curves in the embedding space. This technique further allows visualization of individual predictions by providing the projected position along the learnt curve.
与外周血不同,骨髓显微镜图像中的细胞不仅具有细胞谱系的特征,而且具有谱系内的成熟阶段。由于成熟是一个连续的过程,不同阶段之间的分化属于(有序)回归的范畴。在这项工作中,我们提出了空间成熟度回归-一种规范学习过程的技术,以强制在嵌入空间中对成熟度阶段进行合理定位。为此,我们提出并评估了几种包含该领域知识的曲线模型、目标定义和损失函数。我们表明,当在嵌入空间中沿可学习曲线实施回归目标时,分类f分数提高了2.4个百分点。这种技术通过提供学习曲线上的投影位置,进一步实现了个体预测的可视化。
{"title":"Spatial Maturity Regression for the Classification of Hematopoietic Cells","authors":"Philipp Gräbel, Julian Thull, M. Crysandt, B. Klinkhammer, P. Boor, T. Brümmendorf, D. Merhof","doi":"10.1109/IPTA54936.2022.9784130","DOIUrl":"https://doi.org/10.1109/IPTA54936.2022.9784130","url":null,"abstract":"In contrast to peripheral blood, cells in bone marrow microscopy images are not only characterized by the cell lineage but also a maturity stage within the lineage. As maturation is a continuous process, the differentiation between various stages falls into the category of (ordinal) regression. In this work, we propose Spatial Maturity Regression - a technique that regularizes the learning process to enforce a sensible positioning of maturity stages in the embedding space. To this end, we propose and evaluate several curve models, target definitions and loss function that incorporate this domain knowledge. We show that the classification F-scores improve up to 2.4 percentage points when enforcing regression targets along learnable curves in the embedding space. This technique further allows visualization of individual predictions by providing the projected position along the learnt curve.","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"78 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126297587","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CNN for Hand Washing Movement Classification: What Matters More - the Approach or the Dataset? CNN洗手动作分类:哪个更重要——方法还是数据集?
Atis Elsts, M. Ivanovs, R. Kadikis, O. Sabelnikovs
Good hand hygiene is one of the key factors in preventing infectious diseases, including COVID-19. Advances in machine learning have enabled automated hand hygiene evaluation, with research papers reporting highly accurate hand washing movement classification from video data. However, existing studies typically use datasets collected in lab conditions. In this paper, we apply state-of-the-art techniques such as MobileNetV2 based CNN, including two-stream and recurrent CNN, to three different datasets: a good-quality and uniform lab-based dataset, a more diverse lab-based dataset, and a large-scale real-life dataset collected in a hospital. The results show that while many of the approaches show good accuracy on the first dataset, the accuracy drops significantly o n t he m ore complex datasets. Moreover, all approaches fail to generalize on the third dataset, and only show slightly-better-than random accuracy on videos held out from the training set. This suggests that despite the high accuracy routinely reported in the research literature, the transition to real-world applications for hand washing quality monitoring is not going to be straightforward.
良好的手部卫生是预防包括COVID-19在内的传染病的关键因素之一。机器学习的进步使自动手部卫生评估成为可能,研究论文报告了从视频数据中高度准确的洗手动作分类。然而,现有的研究通常使用在实验室条件下收集的数据集。在本文中,我们将最先进的技术,如基于MobileNetV2的CNN,包括两流和循环CNN,应用于三个不同的数据集:一个高质量和统一的实验室数据集,一个更多样化的实验室数据集,以及一个在医院收集的大规模真实数据集。结果表明,虽然许多方法在第一个数据集上显示出良好的准确性,但在更复杂的数据集上,准确性显着下降。此外,所有的方法都不能在第三个数据集上泛化,并且只在训练集的视频上显示出略好于随机的准确性。这表明,尽管研究文献中经常报道的准确度很高,但将洗手质量监测过渡到现实世界的应用并不简单。
{"title":"CNN for Hand Washing Movement Classification: What Matters More - the Approach or the Dataset?","authors":"Atis Elsts, M. Ivanovs, R. Kadikis, O. Sabelnikovs","doi":"10.1109/IPTA54936.2022.9784153","DOIUrl":"https://doi.org/10.1109/IPTA54936.2022.9784153","url":null,"abstract":"Good hand hygiene is one of the key factors in preventing infectious diseases, including COVID-19. Advances in machine learning have enabled automated hand hygiene evaluation, with research papers reporting highly accurate hand washing movement classification from video data. However, existing studies typically use datasets collected in lab conditions. In this paper, we apply state-of-the-art techniques such as MobileNetV2 based CNN, including two-stream and recurrent CNN, to three different datasets: a good-quality and uniform lab-based dataset, a more diverse lab-based dataset, and a large-scale real-life dataset collected in a hospital. The results show that while many of the approaches show good accuracy on the first dataset, the accuracy drops significantly o n t he m ore complex datasets. Moreover, all approaches fail to generalize on the third dataset, and only show slightly-better-than random accuracy on videos held out from the training set. This suggests that despite the high accuracy routinely reported in the research literature, the transition to real-world applications for hand washing quality monitoring is not going to be straightforward.","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132875635","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Oral Session 1 口头会议1
François Brémond
Message from Program Chairs.................................................................................................................xii Conference Organization..........................................................................................................................xiv Reviewers...................................................................................................................................................xvi
消息从程序的椅子 .................................................................................................................十二会议组织 ..........................................................................................................................十四评论者 ................................................................................................................................................... 十六世
{"title":"Oral Session 1","authors":"François Brémond","doi":"10.1109/ipta54936.2022.9784148","DOIUrl":"https://doi.org/10.1109/ipta54936.2022.9784148","url":null,"abstract":"Message from Program Chairs.................................................................................................................xii Conference Organization..........................................................................................................................xiv Reviewers...................................................................................................................................................xvi","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"78 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132267176","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Analyzing the Variability of Subjective Image Quality Ratings for Different Distortions 分析不同失真情况下主观图像质量评分的可变性
Olga Cherepkova, S. A. Amirshahi, Marius Pedersen
When it comes to evaluating the quality of images, individual observers show different opinions depending on the type of distortion affecting the quality of the image. While in the opinion of one observer a distortion could have a dramatic influence on the quality of the image, another observer could see the same distortion as not having an important effect on the quality of the same image. Using a subjective experiment, we aim to identify the distortions which show the largest variability among observers. For this, 22 observers evaluated the quality of 10 reference images and the 630 test images created from them (21 distortions at three levels). Our results show that the highest variability in subjective scores is linked to distortions like saturation, contrast, sharpness, quantization, some types of added noise, and radial lens distortion.
当谈到评价图像质量时,不同的观察者会根据影响图像质量的失真类型表现出不同的观点。虽然在一个观察者看来,一种扭曲可能对图像的质量产生巨大影响,但另一个观察者可能认为,同样的扭曲对同一图像的质量没有重大影响。使用主观实验,我们的目标是确定在观察者之间显示最大变异性的扭曲。为此,22名观察员评估了10个参考图像和630个测试图像的质量(21个三级失真)。我们的研究结果表明,主观评分的最高可变性与饱和度、对比度、清晰度、量化、某些类型的附加噪声和径向透镜畸变等失真有关。
{"title":"Analyzing the Variability of Subjective Image Quality Ratings for Different Distortions","authors":"Olga Cherepkova, S. A. Amirshahi, Marius Pedersen","doi":"10.1109/IPTA54936.2022.9784120","DOIUrl":"https://doi.org/10.1109/IPTA54936.2022.9784120","url":null,"abstract":"When it comes to evaluating the quality of images, individual observers show different opinions depending on the type of distortion affecting the quality of the image. While in the opinion of one observer a distortion could have a dramatic influence on the quality of the image, another observer could see the same distortion as not having an important effect on the quality of the same image. Using a subjective experiment, we aim to identify the distortions which show the largest variability among observers. For this, 22 observers evaluated the quality of 10 reference images and the 630 test images created from them (21 distortions at three levels). Our results show that the highest variability in subjective scores is linked to distortions like saturation, contrast, sharpness, quantization, some types of added noise, and radial lens distortion.","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125845044","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Special Session 2: Explainable AI for Medical Imaging 特别会议2:可解释的医学成像人工智能
{"title":"Special Session 2: Explainable AI for Medical Imaging","authors":"","doi":"10.1109/ipta54936.2022.9784137","DOIUrl":"https://doi.org/10.1109/ipta54936.2022.9784137","url":null,"abstract":"","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"179 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115480200","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Explainability for Medical Image Captioning 医学图像字幕的可解释性
D. Beddiar, Mourad Oussalah, T. Seppänen
Medical image captioning is the process of generating clinically significant descriptions to medical images, which has many applications among which medical report generation is the most frequent one. In general, automatic captioning of medical images is of great interest for medical experts since it offers assistance in diagnosis, disease treatment and automating the workflow of the health practitioners. Recently, many efforts have been put forward to obtain accurate descriptions but medical image captioning still provides weak and incorrect descriptions. To alleviate this issue, it is important to explain why the model produced a particular caption based on some specific features. This is performed through Artificial Intelligence Explainability (XAI), which aims to unfold the ‘black-box’ feature of deep-learning based models. We present in this paper an explainable module for medical image captioning that provides a sound interpretation of our attention-based encoder-decoder model by explaining the correspondence between visual features and semantic features. We exploit for that, self-attention to compute word importance of semantic features and visual attention to compute relevant regions of the image that correspond to each generated word of the caption in addition to visualization of visual features extracted at each layer of the Convolutional Neural Network (CNN) encoder. We finally evaluate our model using the ImageCLEF medical captioning dataset.
医学图像字幕是对医学图像生成具有临床意义的描述的过程,有许多应用,其中医学报告生成是最常见的一种。一般来说,医学图像的自动字幕是医学专家非常感兴趣的,因为它为诊断、疾病治疗和健康从业者的自动化工作流程提供了帮助。近年来,为了获得准确的描述,人们做出了许多努力,但医学图像字幕仍然提供了较弱和不正确的描述。为了缓解这个问题,解释为什么模型根据一些特定的特征产生特定的标题是很重要的。这是通过人工智能可解释性(XAI)来实现的,该技术旨在揭示基于深度学习的模型的“黑箱”特征。我们在本文中提出了一个可解释的医学图像字幕模块,通过解释视觉特征和语义特征之间的对应关系,为我们基于注意力的编码器-解码器模型提供了一个合理的解释。为此,我们利用自注意来计算语义特征的单词重要性,并利用视觉注意来计算图像中与标题中每个生成的单词对应的相关区域,此外还利用卷积神经网络(CNN)编码器的每一层提取的视觉特征进行可视化。最后,我们使用ImageCLEF医疗字幕数据集评估我们的模型。
{"title":"Explainability for Medical Image Captioning","authors":"D. Beddiar, Mourad Oussalah, T. Seppänen","doi":"10.1109/IPTA54936.2022.9784146","DOIUrl":"https://doi.org/10.1109/IPTA54936.2022.9784146","url":null,"abstract":"Medical image captioning is the process of generating clinically significant descriptions to medical images, which has many applications among which medical report generation is the most frequent one. In general, automatic captioning of medical images is of great interest for medical experts since it offers assistance in diagnosis, disease treatment and automating the workflow of the health practitioners. Recently, many efforts have been put forward to obtain accurate descriptions but medical image captioning still provides weak and incorrect descriptions. To alleviate this issue, it is important to explain why the model produced a particular caption based on some specific features. This is performed through Artificial Intelligence Explainability (XAI), which aims to unfold the ‘black-box’ feature of deep-learning based models. We present in this paper an explainable module for medical image captioning that provides a sound interpretation of our attention-based encoder-decoder model by explaining the correspondence between visual features and semantic features. We exploit for that, self-attention to compute word importance of semantic features and visual attention to compute relevant regions of the image that correspond to each generated word of the caption in addition to visualization of visual features extracted at each layer of the Convolutional Neural Network (CNN) encoder. We finally evaluate our model using the ImageCLEF medical captioning dataset.","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115825261","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Novel and Fast Approach for Reconstructing CASSI-Raman Spectra using Generative Adversarial Networks 一种基于生成对抗网络的CASSI-Raman光谱重构新方法
Jacob Eek, David Gustafsson, Ludwig Hollmann, M. Nordberg, I. Skog, Magnus Malmström
Raman spectroscopy in conjunction with a Coded Aperture Snapshot Spectral Imaging (CASSI) system allows for detection of small amounts of explosives from stand-off distances. The obtained Compressed Sensing (CS) measurements from CASSI consists of mixed spatial and spectral information, from which a HyperSpectral Image (HSI) can be reconstructed. The HSI contains Raman spectra for all spatial locations in the scene, revealing the existence of substances. In this paper we present the possibility of utilizing a learned prior in the form of a conditional generative model for HSI reconstruction using CS. A Generative Adversarial Network (GAN) is trained using simulated samples of HSI, and conditioning on their respective CASSI measurements to refine the prior. Two different types of simulated HSI were investigated, where spatial overlap of substances was either allowed or disallowed. The results show that the developed method produces precise reconstructions of HSI from their CASSI measurements in a matter of seconds.
拉曼光谱与编码孔径快照光谱成像(CASSI)系统相结合,可以从隔离距离检测少量爆炸物。CASSI压缩感知(CS)测量结果由混合的空间和光谱信息组成,可以重建高光谱图像(HSI)。HSI包含场景中所有空间位置的拉曼光谱,揭示物质的存在。在本文中,我们提出了利用条件生成模型形式的学习先验的可能性,用于使用CS进行HSI重建。生成对抗网络(GAN)使用HSI的模拟样本进行训练,并对其各自的CASSI测量进行调节以改进先验。研究了两种不同类型的模拟HSI,其中允许或不允许物质的空间重叠。结果表明,开发的方法可以在几秒钟内从他们的CASSI测量结果中产生精确的HSI重建。
{"title":"A Novel and Fast Approach for Reconstructing CASSI-Raman Spectra using Generative Adversarial Networks","authors":"Jacob Eek, David Gustafsson, Ludwig Hollmann, M. Nordberg, I. Skog, Magnus Malmström","doi":"10.1109/IPTA54936.2022.9784152","DOIUrl":"https://doi.org/10.1109/IPTA54936.2022.9784152","url":null,"abstract":"Raman spectroscopy in conjunction with a Coded Aperture Snapshot Spectral Imaging (CASSI) system allows for detection of small amounts of explosives from stand-off distances. The obtained Compressed Sensing (CS) measurements from CASSI consists of mixed spatial and spectral information, from which a HyperSpectral Image (HSI) can be reconstructed. The HSI contains Raman spectra for all spatial locations in the scene, revealing the existence of substances. In this paper we present the possibility of utilizing a learned prior in the form of a conditional generative model for HSI reconstruction using CS. A Generative Adversarial Network (GAN) is trained using simulated samples of HSI, and conditioning on their respective CASSI measurements to refine the prior. Two different types of simulated HSI were investigated, where spatial overlap of substances was either allowed or disallowed. The results show that the developed method produces precise reconstructions of HSI from their CASSI measurements in a matter of seconds.","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128979204","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Special Session 4: Network Science in Computer Vision 专题会议4:计算机视觉中的网络科学
{"title":"Special Session 4: Network Science in Computer Vision","authors":"","doi":"10.1109/ipta54936.2022.9784145","DOIUrl":"https://doi.org/10.1109/ipta54936.2022.9784145","url":null,"abstract":"","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"323 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133957443","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluation of Multi-Scale Multiple Instance Learning to Improve Thyroid Cancer Classification 多尺度多实例学习改进甲状腺癌分类的评价
M. Tschuchnig, Philipp Grubmüller, Lea Maria Stangassinger, Christina Kreutzer, S. Couillard-Després, G. Oostingh, A. Hittmair, M. Gadermayr
Thyroid cancer is currently the fifth most common malignancy diagnosed in women. Since differentiation of cancer sub-types is important for treatment and current, manual methods are time consuming and subjective, automatic computer-aided differentiation of cancer types is crucial. Manual differentiation of thyroid cancer is based on tissue sections, analysed by pathologists using histological features. Due to the enormous size of gigapixel whole slide images, holistic classification u sing deep learning methods is not feasible. Patch based multiple instance learning approaches, combined with aggre-gations such as bag-of-words, is a common approach. This work's contribution is to extend a patch based state-of-the-art method by generating and combining feature vectors of three different patch resolutions and analysing three distinct ways of combining them. The results showed improvements in one of the three multi-scale approaches, while the others led to decreased scores. This provides motivation for analysis and discussion of the individual approaches.
甲状腺癌是目前女性第五大常见恶性肿瘤。由于癌症亚型的鉴别对治疗很重要,目前,人工方法耗时且主观,因此计算机辅助的癌症类型自动鉴别至关重要。甲状腺癌的人工鉴别是基于组织切片,由病理学家利用组织学特征进行分析。由于十亿像素整张幻灯片图像的巨大尺寸,使用深度学习方法进行整体分类是不可行的。基于Patch的多实例学习方法,结合单词袋等聚合方法,是一种常见的方法。这项工作的贡献是通过生成和组合三种不同补丁分辨率的特征向量并分析三种不同的组合方式来扩展基于最先进的补丁方法。结果显示,三种多尺度方法中的一种方法有所改善,而其他方法则导致得分下降。这为分析和讨论各个方法提供了动力。
{"title":"Evaluation of Multi-Scale Multiple Instance Learning to Improve Thyroid Cancer Classification","authors":"M. Tschuchnig, Philipp Grubmüller, Lea Maria Stangassinger, Christina Kreutzer, S. Couillard-Després, G. Oostingh, A. Hittmair, M. Gadermayr","doi":"10.48550/arXiv.2204.10942","DOIUrl":"https://doi.org/10.48550/arXiv.2204.10942","url":null,"abstract":"Thyroid cancer is currently the fifth most common malignancy diagnosed in women. Since differentiation of cancer sub-types is important for treatment and current, manual methods are time consuming and subjective, automatic computer-aided differentiation of cancer types is crucial. Manual differentiation of thyroid cancer is based on tissue sections, analysed by pathologists using histological features. Due to the enormous size of gigapixel whole slide images, holistic classification u sing deep learning methods is not feasible. Patch based multiple instance learning approaches, combined with aggre-gations such as bag-of-words, is a common approach. This work's contribution is to extend a patch based state-of-the-art method by generating and combining feature vectors of three different patch resolutions and analysing three distinct ways of combining them. The results showed improvements in one of the three multi-scale approaches, while the others led to decreased scores. This provides motivation for analysis and discussion of the individual approaches.","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132997593","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1