首页 > 最新文献

Proceedings of the 2023 Asia Conference on Computer Vision, Image Processing and Pattern Recognition最新文献

英文 中文
Medical Image Segmentation Approach via Transformer Knowledge Distillation 基于变压器知识蒸馏的医学图像分割方法
Tianshu Zhang, Hao Wang, K. Lam, Chi-Yin Chow
Numerous transformer-based medical image segmentation methods have been proposed and achieved good segmentation results. However, it is still a challenge to train and deploy transformer networks to mobile medical devices due to a large number of model parameters. To resolve the training and model parameter problems, in this paper, we propose a Transformer-based network for Medical Image Segmentation using Knowledge Distillation named MISTKD. The MISTKD consists of a teacher network and a student network. It achieves comparable performance to state-of-the-art transformer works using fewer parameters by employing the teacher network to train the student network. The training can be implemented by extracting the sequence in the teacher and student encoder networks during the training procedure. The losses between sequences are further calculated, thus the student network can learn from the teacher network. The experimental results on Synapse show that the proposed work achieves competitive performance using only one-eighth parameters.
人们提出了许多基于变换的医学图像分割方法,并取得了良好的分割效果。然而,由于大量的模型参数,对移动医疗设备的变压器网络进行训练和部署仍然是一个挑战。为了解决训练和模型参数问题,本文提出了一种基于变压器的医学图像分割网络——MISTKD。MISTKD由一个教师网络和一个学生网络组成。它通过使用教师网络来训练学生网络,以更少的参数实现了与最先进的变压器工程相当的性能。训练可以通过在训练过程中提取教师和学生编码器网络中的序列来实现。进一步计算序列之间的损失,从而使学生网络可以从教师网络中学习。在Synapse上的实验结果表明,所提出的工作仅使用八分之一的参数就达到了竞争性能。
{"title":"Medical Image Segmentation Approach via Transformer Knowledge Distillation","authors":"Tianshu Zhang, Hao Wang, K. Lam, Chi-Yin Chow","doi":"10.1145/3596286.3596292","DOIUrl":"https://doi.org/10.1145/3596286.3596292","url":null,"abstract":"Numerous transformer-based medical image segmentation methods have been proposed and achieved good segmentation results. However, it is still a challenge to train and deploy transformer networks to mobile medical devices due to a large number of model parameters. To resolve the training and model parameter problems, in this paper, we propose a Transformer-based network for Medical Image Segmentation using Knowledge Distillation named MISTKD. The MISTKD consists of a teacher network and a student network. It achieves comparable performance to state-of-the-art transformer works using fewer parameters by employing the teacher network to train the student network. The training can be implemented by extracting the sequence in the teacher and student encoder networks during the training procedure. The losses between sequences are further calculated, thus the student network can learn from the teacher network. The experimental results on Synapse show that the proposed work achieves competitive performance using only one-eighth parameters.","PeriodicalId":208318,"journal":{"name":"Proceedings of the 2023 Asia Conference on Computer Vision, Image Processing and Pattern Recognition","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121873756","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Remaining useful life prediction via K-means clustering analysis and deep convolutional neural network 基于k均值聚类分析和深度卷积神经网络的剩余使用寿命预测
Yuru Zhang, Chun-Ming Su, Jiajun Wu
To improve the prediction accuracy of remaining useful life (RUL), a deep learning method coupled with clustering analysis is proposed. K-means clustering algorithm is employed to analyze the operation settings in data set for matching different operating conditions, and a wise operation mechanism is utilized to normalize the sensor data and match the operation history corresponding to the time instances. The deep convolutional neural network (DCNN) architecture is constructed, which adopts time-sliding window-based sequence as network input. Moreover, it does not require expertise in prediction and signal processing. The CMAPSS dataset published by NASA is used for case study. The proposed approach is validated by comparing with other approaches. The results indicate its superiority on prediction performance of RUL for aeroengine.
为了提高剩余使用寿命的预测精度,提出了一种结合聚类分析的深度学习方法。采用K-means聚类算法对数据集中的运行设置进行分析,匹配不同的运行工况,并采用明智的运行机制对传感器数据进行归一化,匹配对应时间实例的运行历史。构造了采用基于时间滑动窗口的序列作为网络输入的深度卷积神经网络(DCNN)体系结构。此外,它不需要预测和信号处理方面的专业知识。案例研究使用NASA发布的CMAPSS数据集。通过与其他方法的比较,验证了该方法的有效性。结果表明,该方法在航空发动机RUL预测性能上具有优越性。
{"title":"Remaining useful life prediction via K-means clustering analysis and deep convolutional neural network","authors":"Yuru Zhang, Chun-Ming Su, Jiajun Wu","doi":"10.1145/3596286.3596297","DOIUrl":"https://doi.org/10.1145/3596286.3596297","url":null,"abstract":"To improve the prediction accuracy of remaining useful life (RUL), a deep learning method coupled with clustering analysis is proposed. K-means clustering algorithm is employed to analyze the operation settings in data set for matching different operating conditions, and a wise operation mechanism is utilized to normalize the sensor data and match the operation history corresponding to the time instances. The deep convolutional neural network (DCNN) architecture is constructed, which adopts time-sliding window-based sequence as network input. Moreover, it does not require expertise in prediction and signal processing. The CMAPSS dataset published by NASA is used for case study. The proposed approach is validated by comparing with other approaches. The results indicate its superiority on prediction performance of RUL for aeroengine.","PeriodicalId":208318,"journal":{"name":"Proceedings of the 2023 Asia Conference on Computer Vision, Image Processing and Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131004010","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Improving Object Detection Robustness against Natural Perturbations through Synthetic Data Augmentation 通过合成数据增强提高目标检测对自然扰动的鲁棒性
N. Premakumara, Brian Jalaian, N. Suri, H. Samani
Robustness against real-world distribution shifts is crucial for the successful deployment of object detection models in practical applications. In this paper, we address the problem of assessing and enhancing the robustness of object detection models against natural perturbations, such as varying lighting conditions, blur, and brightness. We analyze four state-of-the-art deep neural network models, Detr-ResNet-101, Detr-ResNet-50, YOLOv4, and YOLOv4-tiny, using the COCO 2017 dataset and ExDark dataset. By simulating synthetic perturbations with the AugLy package, we systematically explore the optimal level of synthetic perturbation required to improve the models’ robustness through data augmentation techniques. Our comprehensive ablation study meticulously evaluates the impact of synthetic perturbations on object detection models’ performance against real-world distribution shifts, establishing a tangible connection between synthetic augmentation and real-world robustness. Our findings not only substantiate the effectiveness of synthetic perturbations in improving model robustness, but also provide valuable insights for researchers and practitioners in developing more robust and reliable object detection models tailored for real-world applications.
对真实世界分布变化的鲁棒性对于在实际应用中成功部署目标检测模型至关重要。在本文中,我们解决了评估和增强目标检测模型对自然扰动的鲁棒性的问题,例如不同的照明条件,模糊和亮度。我们使用COCO 2017数据集和ExDark数据集分析了四种最先进的深度神经网络模型,即Detr-ResNet-101、Detr-ResNet-50、YOLOv4和YOLOv4-tiny。通过使用AugLy包模拟合成扰动,我们系统地探索了通过数据增强技术提高模型鲁棒性所需的最佳合成扰动水平。我们的综合消融研究细致地评估了合成扰动对目标检测模型在现实世界分布变化下的性能的影响,在合成增强和现实世界鲁棒性之间建立了切实的联系。我们的研究结果不仅证实了综合扰动在提高模型鲁棒性方面的有效性,而且为研究人员和从业者开发更鲁棒、更可靠的目标检测模型提供了有价值的见解。
{"title":"Improving Object Detection Robustness against Natural Perturbations through Synthetic Data Augmentation","authors":"N. Premakumara, Brian Jalaian, N. Suri, H. Samani","doi":"10.1145/3596286.3596293","DOIUrl":"https://doi.org/10.1145/3596286.3596293","url":null,"abstract":"Robustness against real-world distribution shifts is crucial for the successful deployment of object detection models in practical applications. In this paper, we address the problem of assessing and enhancing the robustness of object detection models against natural perturbations, such as varying lighting conditions, blur, and brightness. We analyze four state-of-the-art deep neural network models, Detr-ResNet-101, Detr-ResNet-50, YOLOv4, and YOLOv4-tiny, using the COCO 2017 dataset and ExDark dataset. By simulating synthetic perturbations with the AugLy package, we systematically explore the optimal level of synthetic perturbation required to improve the models’ robustness through data augmentation techniques. Our comprehensive ablation study meticulously evaluates the impact of synthetic perturbations on object detection models’ performance against real-world distribution shifts, establishing a tangible connection between synthetic augmentation and real-world robustness. Our findings not only substantiate the effectiveness of synthetic perturbations in improving model robustness, but also provide valuable insights for researchers and practitioners in developing more robust and reliable object detection models tailored for real-world applications.","PeriodicalId":208318,"journal":{"name":"Proceedings of the 2023 Asia Conference on Computer Vision, Image Processing and Pattern Recognition","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133466973","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Self Attention in U-Net for Semantic Segmentation of Low Resolution SAR Images 低分辨率SAR图像语义分割的U-Net自关注
Hrishikesh Singh Yadav, Priyanshu Panchal, Divyanshu Manawat, G. S, S. S
The SAR image semantic segmentation using computer vision techniques has gained much popularity in the research community due to their wide applications. Despite the advancements in Deep Learning for image analysis, these models still struggle to segment SAR images due to the existence of speckle noise and a poor feature extractor. Moreover, deep learning models are challenging to train on small datasets and the performance of the model is significantly impacted by the quality of the data. This calls for the development of an effective network that can draw out critical information from the low resolution SAR images. In this regard, the present work proposes a unique Self attention module in U-Net for the semantic segmentation of low resolution SAR images.. The Self Attention Model makes use of Laplacian kernel to highlight the sharp discontinuities in the features that define the boundaries of the objects. The proposed model, employs dilated convolution layers at the initial layers, enabling the model to more effectively capture larger contextual information. With an accuracy of 0.84 and an F1-score of 0.83, the proposed model outperforms the state-of-the-art techniques in semantic segmentation of low resolution SAR images. The results clearly demonstrate the importance of the self attention module and the consideration of dilated convolution layers in the initial layers in semantic segmentation of low resolution SAR images.
基于计算机视觉的SAR图像语义分割技术因其广泛的应用而受到了研究界的广泛关注。尽管深度学习在图像分析方面取得了进步,但由于存在散斑噪声和较差的特征提取器,这些模型仍然难以分割SAR图像。此外,在小数据集上训练深度学习模型具有挑战性,并且模型的性能受到数据质量的显着影响。这就要求开发一个有效的网络,从低分辨率SAR图像中提取关键信息。在这方面,本研究在U-Net中提出了一个独特的自关注模块,用于低分辨率SAR图像的语义分割。自注意模型利用拉普拉斯核来突出定义物体边界的特征中的明显不连续。该模型在初始层采用扩展卷积层,使模型能够更有效地捕获更大的上下文信息。该模型的精度为0.84,f1分数为0.83,在低分辨率SAR图像的语义分割中优于最先进的技术。结果表明,在低分辨率SAR图像的语义分割中,自关注模块的重要性和初始层中扩展卷积层的考虑。
{"title":"Self Attention in U-Net for Semantic Segmentation of Low Resolution SAR Images","authors":"Hrishikesh Singh Yadav, Priyanshu Panchal, Divyanshu Manawat, G. S, S. S","doi":"10.1145/3596286.3596291","DOIUrl":"https://doi.org/10.1145/3596286.3596291","url":null,"abstract":"The SAR image semantic segmentation using computer vision techniques has gained much popularity in the research community due to their wide applications. Despite the advancements in Deep Learning for image analysis, these models still struggle to segment SAR images due to the existence of speckle noise and a poor feature extractor. Moreover, deep learning models are challenging to train on small datasets and the performance of the model is significantly impacted by the quality of the data. This calls for the development of an effective network that can draw out critical information from the low resolution SAR images. In this regard, the present work proposes a unique Self attention module in U-Net for the semantic segmentation of low resolution SAR images.. The Self Attention Model makes use of Laplacian kernel to highlight the sharp discontinuities in the features that define the boundaries of the objects. The proposed model, employs dilated convolution layers at the initial layers, enabling the model to more effectively capture larger contextual information. With an accuracy of 0.84 and an F1-score of 0.83, the proposed model outperforms the state-of-the-art techniques in semantic segmentation of low resolution SAR images. The results clearly demonstrate the importance of the self attention module and the consideration of dilated convolution layers in the initial layers in semantic segmentation of low resolution SAR images.","PeriodicalId":208318,"journal":{"name":"Proceedings of the 2023 Asia Conference on Computer Vision, Image Processing and Pattern Recognition","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124950825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Proceedings of the 2023 Asia Conference on Computer Vision, Image Processing and Pattern Recognition 2023亚洲计算机视觉、图像处理与模式识别会议论文集
{"title":"Proceedings of the 2023 Asia Conference on Computer Vision, Image Processing and Pattern Recognition","authors":"","doi":"10.1145/3596286","DOIUrl":"https://doi.org/10.1145/3596286","url":null,"abstract":"","PeriodicalId":208318,"journal":{"name":"Proceedings of the 2023 Asia Conference on Computer Vision, Image Processing and Pattern Recognition","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127051015","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the 2023 Asia Conference on Computer Vision, Image Processing and Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1