首页 > 最新文献

2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)最新文献

英文 中文
Special Session 5: Processing and Protection of Encrypted Multimedia Data 专题会议5:加密多媒体数据的处理和保护
{"title":"Special Session 5: Processing and Protection of Encrypted Multimedia Data","authors":"","doi":"10.1109/ipta54936.2022.9784117","DOIUrl":"https://doi.org/10.1109/ipta54936.2022.9784117","url":null,"abstract":"","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"2008 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125579928","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Ψ-NET: A Novel Encoder-Decoder Architecture for Animal Segmentation Ψ-NET:一种用于动物分割的新型编码器-解码器架构
David Norman Díaz Estrada, Utkarsh Goyal, M. Ullah, F. A. Cheikh
This paper proposes a novel Ψ-Net architecture that consists of three encoders and a decoder for animal image segmentation. The main characteristic of our proposed architecture is that the outputs at each depth level of the three encoders are summed up and then concatenated in the corresponding depth levels of the decoder for the upsampling process. We col-lected a new dataset consisting of 200 images for training the model, and we manually labelled the ground truth segmentation masks for these images. We trained our proposed model Ψ-Net on this dataset and compared the segmentation accu-racy with the classical U-Net and Y-Net architectures. Our proposed model achieved the highest accuracy on the dataset with 93% pixel accuracy, and 81.6% mean intersection-over-union (IoU).
本文提出了一种新颖的Ψ-Net架构,该架构由三个编码器和一个解码器组成,用于动物图像分割。我们提出的架构的主要特点是,三个编码器的每个深度级别的输出被汇总,然后在解码器的相应深度级别中进行上采样过程的连接。我们收集了一个由200张图像组成的新数据集来训练模型,并为这些图像手动标记了地面真值分割掩码。我们在该数据集上训练了我们提出的模型Ψ-Net,并将分割精度与经典的U-Net和Y-Net架构进行了比较。我们提出的模型在数据集上达到了最高的精度,像素精度为93%,平均交叉-超联合(IoU)为81.6%。
{"title":"Ψ-NET: A Novel Encoder-Decoder Architecture for Animal Segmentation","authors":"David Norman Díaz Estrada, Utkarsh Goyal, M. Ullah, F. A. Cheikh","doi":"10.1109/IPTA54936.2022.9784135","DOIUrl":"https://doi.org/10.1109/IPTA54936.2022.9784135","url":null,"abstract":"This paper proposes a novel Ψ-Net architecture that consists of three encoders and a decoder for animal image segmentation. The main characteristic of our proposed architecture is that the outputs at each depth level of the three encoders are summed up and then concatenated in the corresponding depth levels of the decoder for the upsampling process. We col-lected a new dataset consisting of 200 images for training the model, and we manually labelled the ground truth segmentation masks for these images. We trained our proposed model Ψ-Net on this dataset and compared the segmentation accu-racy with the classical U-Net and Y-Net architectures. Our proposed model achieved the highest accuracy on the dataset with 93% pixel accuracy, and 81.6% mean intersection-over-union (IoU).","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121928437","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Monoplanar CT Reconstruction with GANs gan的单平面CT重建
Justus Schock, Yu-Chia Lan, D. Truhn, M. Kopaczka, Stefan Conrad, S. Nebelung, D. Merhof
Reconstructing Computed Tomography images (CT) from radiographs currently requires biplanar radiographs for accurate CT reconstruction due to the complementary information contained in the individual views. However, in many cases biplanar information is not available. In this work, we therefore propose a KNN and a PCA-based approach using biplanar radiographs only at the training stage while performing the final inference using only a single anterior-posterior radiograph, thereby increasing the applicability and usability of the model. The methods are quantitatively validated on a multiview database achieving 81% PSNR of biplanar inference and also qualitatively on a dataset of radiographs with no corresponding CT scans.
从x线片重建计算机断层扫描图像(CT)目前需要双平面x线片进行精确的CT重建,因为单个视图中包含互补信息。然而,在许多情况下,双平面信息是不可用的。因此,在这项工作中,我们提出了一种KNN和基于pca的方法,仅在训练阶段使用双平面x线片,同时仅使用单个前后x线片进行最终推断,从而提高了模型的适用性和可用性。这些方法在多视图数据库上进行了定量验证,双平面推断的PSNR达到81%,在没有相应CT扫描的x线片数据集上也进行了定性验证。
{"title":"Monoplanar CT Reconstruction with GANs","authors":"Justus Schock, Yu-Chia Lan, D. Truhn, M. Kopaczka, Stefan Conrad, S. Nebelung, D. Merhof","doi":"10.1109/IPTA54936.2022.9784126","DOIUrl":"https://doi.org/10.1109/IPTA54936.2022.9784126","url":null,"abstract":"Reconstructing Computed Tomography images (CT) from radiographs currently requires biplanar radiographs for accurate CT reconstruction due to the complementary information contained in the individual views. However, in many cases biplanar information is not available. In this work, we therefore propose a KNN and a PCA-based approach using biplanar radiographs only at the training stage while performing the final inference using only a single anterior-posterior radiograph, thereby increasing the applicability and usability of the model. The methods are quantitatively validated on a multiview database achieving 81% PSNR of biplanar inference and also qualitatively on a dataset of radiographs with no corresponding CT scans.","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124064473","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Weak supervision using cell tracking annotation and image registration improves cell segmentation 使用细胞跟踪注释和图像配准的弱监督改进了细胞分割
N. A. Anoshina, D. Sorokin
Learning-based cell segmentation methods have proved to be very effective in cell tracking. The main difficulty of using machine learning is the lack of expert annotation of biomedical data. We propose a weakly-supervised approach that extends the amount of segmentation training data for image sequences where only a couple of frames are annotated. The approach uses the tracking annotations as weak labels and image registration to extend the segmentation annotation to the neighbouring frames. This technique was applied to cell segmentation step in the cell tracking problem. An experimental comparison of the baseline segmentation network trained on the data with pure GT annotation and the same segmentation network trained on the GT data and additional annotations generated with the proposed approach has been performed. The proposed weakly-supervised approach increased the IoU and SEG metrics on the data from the Cell Tracking Challenge.
基于学习的细胞分割方法已被证明是非常有效的细胞跟踪方法。使用机器学习的主要困难是缺乏对生物医学数据的专家注释。我们提出了一种弱监督的方法,该方法扩展了只有几个帧被注释的图像序列的分割训练数据量。该方法使用跟踪标注作为弱标签,并使用图像配准将分割标注扩展到相邻帧。将该技术应用于细胞跟踪问题中的细胞分割步骤。实验比较了在纯GT标注数据上训练的基线分割网络和在使用该方法生成的GT数据和附加标注上训练的相同分割网络。提出的弱监督方法提高了细胞跟踪挑战数据的IoU和SEG指标。
{"title":"Weak supervision using cell tracking annotation and image registration improves cell segmentation","authors":"N. A. Anoshina, D. Sorokin","doi":"10.1109/IPTA54936.2022.9784140","DOIUrl":"https://doi.org/10.1109/IPTA54936.2022.9784140","url":null,"abstract":"Learning-based cell segmentation methods have proved to be very effective in cell tracking. The main difficulty of using machine learning is the lack of expert annotation of biomedical data. We propose a weakly-supervised approach that extends the amount of segmentation training data for image sequences where only a couple of frames are annotated. The approach uses the tracking annotations as weak labels and image registration to extend the segmentation annotation to the neighbouring frames. This technique was applied to cell segmentation step in the cell tracking problem. An experimental comparison of the baseline segmentation network trained on the data with pure GT annotation and the same segmentation network trained on the GT data and additional annotations generated with the proposed approach has been performed. The proposed weakly-supervised approach increased the IoU and SEG metrics on the data from the Cell Tracking Challenge.","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134315837","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pyramid Tokens-to-Token Vision Transformer for Thyroid Pathology Image Classification 用于甲状腺病理图像分类的金字塔标记到标记视觉转换器
Peng Yin, Bo Yu, Cheng-wei Jiang, Hechang Chen
Histopathological image contains rich phenotypic information, which is beneficial to classifying tumor subtypes and predicting the development of diseases. The vast size of pathological slides makes it impossible to directly train whole slide images (WSI) on convolutional neural networks (CNNs). Most of the previous weakly supervision works divide high-resolution WSIs into small image patches and separately input them into the CNN to classify them as tumors or normal areas. The first difficulty is that although the method based on the CNN framework achieves a high accuracy rate, it increases the model parameters and computational complexity. The second difficulty is balancing the relationship between accuracy and model compu-tation. It makes the model maintain and improve the classification accuracy as much as possible based on the lightweight. In this paper, we propose a new lightweight architecture called Pyramid Tokens-to-Token VIsion Transformer (PyT2T-ViT) with multiple instance learning based on Vision Transformer. We introduce the feature extractor of the model with Token-to-Token ViT (T2T-ViT) to reduce the model parameters. The performance of the model is improved by combining the image pyramid of multiple receptive fields so that it can take into account the local and global features of the cell structure at a single scale. We applied the method to our collection of 560 thyroid pathology images from the same institution, model parameters and computation were greatly reduced. The classification effect is significantly better than the CNN-based method.
组织病理图像包含丰富的表型信息,有利于肿瘤亚型的分类和疾病发展的预测。病理切片的巨大尺寸使得在卷积神经网络(cnn)上直接训练整个切片图像(WSI)成为不可能。以往的弱监督工作大多是将高分辨率wsi分割成小块图像,分别输入到CNN中进行肿瘤或正常区域的分类。第一个困难是基于CNN框架的方法虽然达到了较高的准确率,但增加了模型参数和计算复杂度。第二个困难是平衡精度和模型计算之间的关系。它使模型在轻量化的基础上尽可能地保持和提高分类精度。在本文中,我们提出了一种新的轻量级架构,称为金字塔令牌到令牌视觉转换器(PyT2T-ViT),它具有基于视觉转换器的多实例学习。我们引入了Token-to-Token ViT (T2T-ViT)模型的特征提取器来减少模型参数。通过结合多个感受野的图像金字塔来提高模型的性能,使其能够在单个尺度上兼顾细胞结构的局部和全局特征。我们将该方法应用于同一机构收集的560张甲状腺病理图像,大大减少了模型参数和计算量。分类效果明显优于基于cnn的方法。
{"title":"Pyramid Tokens-to-Token Vision Transformer for Thyroid Pathology Image Classification","authors":"Peng Yin, Bo Yu, Cheng-wei Jiang, Hechang Chen","doi":"10.1109/IPTA54936.2022.9784139","DOIUrl":"https://doi.org/10.1109/IPTA54936.2022.9784139","url":null,"abstract":"Histopathological image contains rich phenotypic information, which is beneficial to classifying tumor subtypes and predicting the development of diseases. The vast size of pathological slides makes it impossible to directly train whole slide images (WSI) on convolutional neural networks (CNNs). Most of the previous weakly supervision works divide high-resolution WSIs into small image patches and separately input them into the CNN to classify them as tumors or normal areas. The first difficulty is that although the method based on the CNN framework achieves a high accuracy rate, it increases the model parameters and computational complexity. The second difficulty is balancing the relationship between accuracy and model compu-tation. It makes the model maintain and improve the classification accuracy as much as possible based on the lightweight. In this paper, we propose a new lightweight architecture called Pyramid Tokens-to-Token VIsion Transformer (PyT2T-ViT) with multiple instance learning based on Vision Transformer. We introduce the feature extractor of the model with Token-to-Token ViT (T2T-ViT) to reduce the model parameters. The performance of the model is improved by combining the image pyramid of multiple receptive fields so that it can take into account the local and global features of the cell structure at a single scale. We applied the method to our collection of 560 thyroid pathology images from the same institution, model parameters and computation were greatly reduced. The classification effect is significantly better than the CNN-based method.","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115703585","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Hand-Based Person Identification using Global and Part-Aware Deep Feature Representation Learning 基于全局和局部感知深度特征表示学习的手部人物识别
Nathanael L. Baisa, Zheheng Jiang, Ritesh Vyas, Bryan Williams, Hossein Rahmani, P. Angelov, Sue Black
In cases of serious crime, including sexual abuse, often the only available information with demonstrated potential for identification is images of the hands. Since this evidence is captured in uncontrolled situations, it is difficult to analyse. As global approaches to feature comparison are limited in this case, it is important to extend to consider local information. In this work, we propose hand-based person identification by learning both global and local deep feature representations. Our proposed method, Global and Part-Aware Network (GPA-Net), creates global and local branches on the conv-layer for learning robust discriminative global and part-level features. For learning the local (part-level) features, we perform uniform partitioning on the conv-layer in both horizontal and vertical directions. We retrieve the parts by conducting a soft partition without explicitly partitioning the images or requiring external cues such as pose estimation. We make extensive evaluations on two large multi-ethnic and publicly available hand datasets, demonstrating that our proposed method significantly outperforms competing approaches.
在包括性虐待在内的严重犯罪案件中,证明有可能识别身份的唯一可用信息往往是手的图像。由于这些证据是在不受控制的情况下获得的,因此很难进行分析。由于在这种情况下进行特征比较的全局方法是有限的,因此扩展到考虑局部信息是很重要的。在这项工作中,我们通过学习全局和局部深度特征表示提出了基于手的人识别。我们提出的方法,全局和部分感知网络(GPA-Net),在卷积层上创建全局和局部分支,用于学习鲁棒判别全局和部分级特征。为了学习局部(部分级)特征,我们在水平和垂直方向上对卷积层进行统一划分。我们通过进行软分区来检索这些部分,而不需要明确地对图像进行分区或需要外部线索(如姿态估计)。我们对两个大型多种族和公开可用的手数据集进行了广泛的评估,证明我们提出的方法显着优于竞争方法。
{"title":"Hand-Based Person Identification using Global and Part-Aware Deep Feature Representation Learning","authors":"Nathanael L. Baisa, Zheheng Jiang, Ritesh Vyas, Bryan Williams, Hossein Rahmani, P. Angelov, Sue Black","doi":"10.1109/IPTA54936.2022.9784133","DOIUrl":"https://doi.org/10.1109/IPTA54936.2022.9784133","url":null,"abstract":"In cases of serious crime, including sexual abuse, often the only available information with demonstrated potential for identification is images of the hands. Since this evidence is captured in uncontrolled situations, it is difficult to analyse. As global approaches to feature comparison are limited in this case, it is important to extend to consider local information. In this work, we propose hand-based person identification by learning both global and local deep feature representations. Our proposed method, Global and Part-Aware Network (GPA-Net), creates global and local branches on the conv-layer for learning robust discriminative global and part-level features. For learning the local (part-level) features, we perform uniform partitioning on the conv-layer in both horizontal and vertical directions. We retrieve the parts by conducting a soft partition without explicitly partitioning the images or requiring external cues such as pose estimation. We make extensive evaluations on two large multi-ethnic and publicly available hand datasets, demonstrating that our proposed method significantly outperforms competing approaches.","PeriodicalId":381729,"journal":{"name":"2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-01-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134023170","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
2022 Eleventh International Conference on Image Processing Theory, Tools and Applications (IPTA)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1