首页 > 最新文献

2022 IEEE 5th International Conference on Image Processing Applications and Systems (IPAS)最新文献

英文 中文
Vehicle Color Identification Framework using Pixel-level Color Estimation from Segmentation Masks of Car Parts 基于像素级颜色估计的汽车零件分割掩模颜色识别框架
Pub Date : 2022-12-05 DOI: 10.1109/IPAS55744.2022.10052969
Klearchos Stavrothanasopoulos, Konstantinos Gkountakos, K. Ioannidis, T. Tsikrika, S. Vrochidis, Y. Kompatsiaris
Color comprises one of the most significant and dominant cues for various applications. As one of the most noticeable and stable attributes of vehicles, color can constitute a valuable key component in several practices of intelligent surveillance systems. In this paper, we propose a deep-learning-based framework that combines semantic segmentation masks with pixels clustering for automatic vehicle color recognition. Different from conventional methods, which usually consider only the features of the vehicle's front side, the proposed algorithm is able for view-independent color identification, which is more effective for the surveillance tasks. To the best of our knowledge, this is the first work that employs semantic segmentation masks along with color clustering for the extraction of the vehicle's color representative parts and the recognition of the dominant color, respectively. To evaluate the performance of the proposed method, we introduce a challenging multi-view dataset of 500 car-related RGB images extending the publicly available DSMLR Car Parts dataset for vehicle parts segmentation. The experiments demonstrate that the proposed approach achieves excellent performance and accurate results reaching an accuracy of 93.06% in the multi-view scenario. To facilitate further research, the evaluation dataset and the pre-trained models will be released at https://github.com/klearchos-stav/vehicle_color_recognition.
颜色是各种应用中最重要和最主要的线索之一。颜色作为车辆最显著和最稳定的属性之一,在智能监控系统的一些实践中可以构成一个有价值的关键组成部分。在本文中,我们提出了一个基于深度学习的框架,该框架将语义分割掩码与像素聚类相结合,用于自动车辆颜色识别。与传统方法通常只考虑车辆正面特征不同,该算法能够实现与视觉无关的颜色识别,更有效地完成监控任务。据我们所知,这是第一个使用语义分割掩码和颜色聚类来分别提取车辆颜色代表部件和识别主色的工作。为了评估该方法的性能,我们引入了一个具有挑战性的多视图数据集,该数据集包含500张与汽车相关的RGB图像,扩展了公开可用的DSMLR汽车部件数据集,用于汽车部件分割。实验表明,该方法在多视图场景下取得了优异的性能和准确的结果,准确率达到93.06%。为了便于进一步研究,评估数据集和预训练模型将在https://github.com/klearchos-stav/vehicle_color_recognition上发布。
{"title":"Vehicle Color Identification Framework using Pixel-level Color Estimation from Segmentation Masks of Car Parts","authors":"Klearchos Stavrothanasopoulos, Konstantinos Gkountakos, K. Ioannidis, T. Tsikrika, S. Vrochidis, Y. Kompatsiaris","doi":"10.1109/IPAS55744.2022.10052969","DOIUrl":"https://doi.org/10.1109/IPAS55744.2022.10052969","url":null,"abstract":"Color comprises one of the most significant and dominant cues for various applications. As one of the most noticeable and stable attributes of vehicles, color can constitute a valuable key component in several practices of intelligent surveillance systems. In this paper, we propose a deep-learning-based framework that combines semantic segmentation masks with pixels clustering for automatic vehicle color recognition. Different from conventional methods, which usually consider only the features of the vehicle's front side, the proposed algorithm is able for view-independent color identification, which is more effective for the surveillance tasks. To the best of our knowledge, this is the first work that employs semantic segmentation masks along with color clustering for the extraction of the vehicle's color representative parts and the recognition of the dominant color, respectively. To evaluate the performance of the proposed method, we introduce a challenging multi-view dataset of 500 car-related RGB images extending the publicly available DSMLR Car Parts dataset for vehicle parts segmentation. The experiments demonstrate that the proposed approach achieves excellent performance and accurate results reaching an accuracy of 93.06% in the multi-view scenario. To facilitate further research, the evaluation dataset and the pre-trained models will be released at https://github.com/klearchos-stav/vehicle_color_recognition.","PeriodicalId":322228,"journal":{"name":"2022 IEEE 5th International Conference on Image Processing Applications and Systems (IPAS)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127958902","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving 3D Point Cloud Reconstruction with Dynamic Tree-Structured Capsules 用动态树状结构胶囊改进三维点云重建
Pub Date : 2022-12-05 DOI: 10.1109/IPAS55744.2022.10052906
Chris Engelhardt, Jakob Mittelberger, David Peer, Sebastian Stabinger, A. Rodríguez-Sánchez
When applying convolutional neural networks to 3D point cloud reconstruction, these do not seem to be able to learn meaningful 2D manifold embeddings, suffer a lack of explainability and are vulnerable to adversarial attacks [20]. Except for the latter, these shortcomings can be overcome with capsule networks. In this work we introduce an auto-encoder based on dynamic tree-structured capsule networks for sparse 3D point clouds with SDA-routing. Our approach preserves the spatial arrangements of the input data and increases the adversarial robustness without introducing additional computational overhead. Our experimental evaluation shows that our architecture outperforms the current state-of-the-art capsule and CNN-based networks.
当将卷积神经网络应用于3D点云重建时,它们似乎无法学习有意义的2D流形嵌入,缺乏可解释性,并且容易受到对抗性攻击[10]。除了后者,这些缺点都可以用胶囊网络克服。本文介绍了一种基于动态树结构胶囊网络的自编码器,用于稀疏三维点云的sda路由。我们的方法保留了输入数据的空间排列,并在不引入额外计算开销的情况下增加了对抗鲁棒性。我们的实验评估表明,我们的架构优于当前最先进的胶囊和基于cnn的网络。
{"title":"Improving 3D Point Cloud Reconstruction with Dynamic Tree-Structured Capsules","authors":"Chris Engelhardt, Jakob Mittelberger, David Peer, Sebastian Stabinger, A. Rodríguez-Sánchez","doi":"10.1109/IPAS55744.2022.10052906","DOIUrl":"https://doi.org/10.1109/IPAS55744.2022.10052906","url":null,"abstract":"When applying convolutional neural networks to 3D point cloud reconstruction, these do not seem to be able to learn meaningful 2D manifold embeddings, suffer a lack of explainability and are vulnerable to adversarial attacks [20]. Except for the latter, these shortcomings can be overcome with capsule networks. In this work we introduce an auto-encoder based on dynamic tree-structured capsule networks for sparse 3D point clouds with SDA-routing. Our approach preserves the spatial arrangements of the input data and increases the adversarial robustness without introducing additional computational overhead. Our experimental evaluation shows that our architecture outperforms the current state-of-the-art capsule and CNN-based networks.","PeriodicalId":322228,"journal":{"name":"2022 IEEE 5th International Conference on Image Processing Applications and Systems (IPAS)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127673960","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Complex Network for Complex Problems: A comparative study of CNN and Complex-valued CNN 复杂问题的复杂网络:CNN与复值CNN的比较研究
Pub Date : 2022-12-05 DOI: 10.1109/IPAS55744.2022.10053060
S. Chatterjee, Pavan Tummala, O. Speck, A. Nürnberger
Neural networks, especially convolutional neural networks (CNN), are one of the most common tools these days used in computer vision. Most of these networks work with real-valued data using real-valued features. Complex-valued convolutional neural networks (CV-CNN) can preserve the algebraic structure of complex-valued input data and have the potential to learn more complex relationships between the input and the ground-truth. Although some comparisons of CNNs and CV-CNNs for different tasks have been performed in the past, a large-scale investigation comparing different models operating on different tasks has not been conducted. Furthermore, because complex features contain both real and imaginary components, CV-CNNs have double the number of trainable parameters as real-valued CNNs in terms of the actual number of trainable parameters. Whether or not the improvements in performance with CV-CNN observed in the past have been because of the complex features or just because of having double the number of trainable parameters has not yet been explored. This paper presents a comparative study of CNN, CNNx2 (CNN with double the number of trainable parameters as the CNN), and CV-CNN. The experiments were performed using seven models for two different tasks - brain tumour classification and segmentation in brain MRIs. The results have revealed that the CV-CNN models outperformed the CNN and CNNx2 models.
神经网络,尤其是卷积神经网络(CNN),是当今计算机视觉中最常用的工具之一。这些网络大多使用实值特征处理实值数据。复值卷积神经网络(CV-CNN)可以保留复值输入数据的代数结构,并具有学习输入与真值之间更复杂关系的潜力。虽然过去已经对cnn和cv - cnn在不同任务下的运行情况进行了一些比较,但还没有对不同模型在不同任务下的运行情况进行大规模的比较研究。此外,由于复杂特征既包含实分量又包含虚分量,cv - cnn在可训练参数的实际数量上是实值cnn的两倍。过去观察到的CV-CNN在性能上的改进是由于复杂的特征还是仅仅因为可训练参数的数量增加了一倍,目前还没有研究。本文对CNN、CNNx2(可训练参数数是CNN的两倍的CNN)和CV-CNN进行了比较研究。实验使用7个模型来完成两个不同的任务——脑部肿瘤的分类和脑核磁共振成像的分割。结果表明,CV-CNN模型优于CNN和CNNx2模型。
{"title":"Complex Network for Complex Problems: A comparative study of CNN and Complex-valued CNN","authors":"S. Chatterjee, Pavan Tummala, O. Speck, A. Nürnberger","doi":"10.1109/IPAS55744.2022.10053060","DOIUrl":"https://doi.org/10.1109/IPAS55744.2022.10053060","url":null,"abstract":"Neural networks, especially convolutional neural networks (CNN), are one of the most common tools these days used in computer vision. Most of these networks work with real-valued data using real-valued features. Complex-valued convolutional neural networks (CV-CNN) can preserve the algebraic structure of complex-valued input data and have the potential to learn more complex relationships between the input and the ground-truth. Although some comparisons of CNNs and CV-CNNs for different tasks have been performed in the past, a large-scale investigation comparing different models operating on different tasks has not been conducted. Furthermore, because complex features contain both real and imaginary components, CV-CNNs have double the number of trainable parameters as real-valued CNNs in terms of the actual number of trainable parameters. Whether or not the improvements in performance with CV-CNN observed in the past have been because of the complex features or just because of having double the number of trainable parameters has not yet been explored. This paper presents a comparative study of CNN, CNNx2 (CNN with double the number of trainable parameters as the CNN), and CV-CNN. The experiments were performed using seven models for two different tasks - brain tumour classification and segmentation in brain MRIs. The results have revealed that the CV-CNN models outperformed the CNN and CNNx2 models.","PeriodicalId":322228,"journal":{"name":"2022 IEEE 5th International Conference on Image Processing Applications and Systems (IPAS)","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125618703","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Cluster Analysis: Unsupervised Classification for Identifying Benign and Malignant Tumors on Whole Slide Image of Prostate Cancer 聚类分析:无监督分类在前列腺癌全切片图像上鉴别良恶性肿瘤
Pub Date : 2022-12-05 DOI: 10.1109/IPAS55744.2022.10052952
Subrata Bhattacharjee, Yeong-Byn Hwang, Rashadul Islam Sumon, H. Rahman, Dong-Woo Hyeon, Damin Moon, Kouayep Sonia Carole, Hee-Cheol Kim, Heung-Kook Choi
Recently, many fields have widely used cluster analysis: psychology, biology, statistics, pattern recognition, information retrieval, machine learning, and data mining. Diagnosis of histopathological images of prostate cancer is one of the routine tasks for pathologists and it is challenging for pathologists to analyze the formation of glands and tumors based on the Gleason grading system. In this study, unsupervised classification has been performed for differentiating malignant (cancerous) from benign (non-cancerous) tumors. Therefore, the unsupervised-based computer-aided diagnosis (CAD) technique would be of great benefit in easing the workloads of pathologists. This technique is used to find meaningful clustering objects (i.e., individuals, entities, patterns, or cases) and identify useful patterns. Radiomic-based features were extracted for cluster analysis using the gray-level co-occurrence matrix (GLCM), gray-level run-length matrix (GLRLM), and gray-level size zone matrix (GLSZM) techniques. Multi-clustering techniques used for the unsupervised classification are K-means clustering, K-medoids clustering, Agglomerative Hierarchical (AH) clustering, Gaussian mixture model (GMM) clustering, and Spectral clustering. The quality of the clustering algorithms was determined using Purity, Silhouettes, Adjusted Rand, Fowlkes Mallows, and Calinski Harabasz (CH) scores. However, the best-performing algorithm (i.e., K-means) has been applied to predict and annotate the cancerous regions in the whole slide image (WSI) to compare with the pathologist annotation.
最近,许多领域广泛使用聚类分析:心理学、生物学、统计学、模式识别、信息检索、机器学习和数据挖掘。前列腺癌组织病理图像的诊断是病理学家的常规任务之一,基于Gleason分级系统分析腺体和肿瘤的形成对病理学家来说是一个挑战。在本研究中,非监督分类被用于区分恶性(癌性)和良性(非癌性)肿瘤。因此,基于无监督的计算机辅助诊断(CAD)技术将极大地减轻病理学家的工作量。该技术用于查找有意义的聚类对象(即,个体、实体、模式或案例)并识别有用的模式。利用灰度共现矩阵(GLCM)、灰度游程矩阵(GLRLM)和灰度大小区域矩阵(GLSZM)技术提取基于放射组学的特征进行聚类分析。用于无监督分类的多聚类技术有K-means聚类、k - medidoids聚类、Agglomerative Hierarchical聚类(AH)聚类、高斯混合模型(GMM)聚类和谱聚类。聚类算法的质量是用Purity、Silhouettes、Adjusted Rand、Fowlkes Mallows和Calinski Harabasz (CH)评分来确定的。然而,在整个幻灯片图像(WSI)中,使用性能最好的算法(即K-means)来预测和注释癌变区域,并与病理学家注释进行比较。
{"title":"Cluster Analysis: Unsupervised Classification for Identifying Benign and Malignant Tumors on Whole Slide Image of Prostate Cancer","authors":"Subrata Bhattacharjee, Yeong-Byn Hwang, Rashadul Islam Sumon, H. Rahman, Dong-Woo Hyeon, Damin Moon, Kouayep Sonia Carole, Hee-Cheol Kim, Heung-Kook Choi","doi":"10.1109/IPAS55744.2022.10052952","DOIUrl":"https://doi.org/10.1109/IPAS55744.2022.10052952","url":null,"abstract":"Recently, many fields have widely used cluster analysis: psychology, biology, statistics, pattern recognition, information retrieval, machine learning, and data mining. Diagnosis of histopathological images of prostate cancer is one of the routine tasks for pathologists and it is challenging for pathologists to analyze the formation of glands and tumors based on the Gleason grading system. In this study, unsupervised classification has been performed for differentiating malignant (cancerous) from benign (non-cancerous) tumors. Therefore, the unsupervised-based computer-aided diagnosis (CAD) technique would be of great benefit in easing the workloads of pathologists. This technique is used to find meaningful clustering objects (i.e., individuals, entities, patterns, or cases) and identify useful patterns. Radiomic-based features were extracted for cluster analysis using the gray-level co-occurrence matrix (GLCM), gray-level run-length matrix (GLRLM), and gray-level size zone matrix (GLSZM) techniques. Multi-clustering techniques used for the unsupervised classification are K-means clustering, K-medoids clustering, Agglomerative Hierarchical (AH) clustering, Gaussian mixture model (GMM) clustering, and Spectral clustering. The quality of the clustering algorithms was determined using Purity, Silhouettes, Adjusted Rand, Fowlkes Mallows, and Calinski Harabasz (CH) scores. However, the best-performing algorithm (i.e., K-means) has been applied to predict and annotate the cancerous regions in the whole slide image (WSI) to compare with the pathologist annotation.","PeriodicalId":322228,"journal":{"name":"2022 IEEE 5th International Conference on Image Processing Applications and Systems (IPAS)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127839643","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A fast method for impulse noise reduction in digital color images using anomaly median filtering 一种利用异常中值滤波快速消除数字彩色图像脉冲噪声的方法
Pub Date : 2022-12-05 DOI: 10.1109/IPAS55744.2022.10052947
S. Gantenapalli, P. Choppala, Vandana Gullipalli, J. Meka, Paul D. Teal
The traditional vector median filtering and its variants used to reduce impulse noise in digital color images operate by processing over all the pixels in the image sequentially. This renders these filtering methods computationally expensive. This paper presents a fast method for reducing impulse noise in digital color images. The key idea here is to slice each row of the image as a univariate data vector, identify impulse noise using anomaly detection schemes and then apply median filtering over these to restore the original image. This idea ensures fast filtering as only the noisy pixels are processed. Using simulations, we show that the proposed method scales efficiently with respect to accuracy and time. Through a combined measure of time and accuracy, we show that the proposed method exhibits nearly 42% improvement over the conventional ones.
传统的矢量中值滤波及其变体用于减少数字彩色图像中的脉冲噪声,是通过对图像中的所有像素进行顺序处理来实现的。这使得这些过滤方法在计算上非常昂贵。提出了一种快速降低数字彩色图像脉冲噪声的方法。这里的关键思想是将图像的每一行切片为单变量数据向量,使用异常检测方案识别脉冲噪声,然后对这些脉冲噪声应用中值滤波以恢复原始图像。这个想法确保快速滤波,因为只有有噪声的像素被处理。通过仿真,我们证明了所提出的方法在精度和时间方面是有效的。通过时间和精度的综合衡量,我们表明该方法比传统方法提高了近42%。
{"title":"A fast method for impulse noise reduction in digital color images using anomaly median filtering","authors":"S. Gantenapalli, P. Choppala, Vandana Gullipalli, J. Meka, Paul D. Teal","doi":"10.1109/IPAS55744.2022.10052947","DOIUrl":"https://doi.org/10.1109/IPAS55744.2022.10052947","url":null,"abstract":"The traditional vector median filtering and its variants used to reduce impulse noise in digital color images operate by processing over all the pixels in the image sequentially. This renders these filtering methods computationally expensive. This paper presents a fast method for reducing impulse noise in digital color images. The key idea here is to slice each row of the image as a univariate data vector, identify impulse noise using anomaly detection schemes and then apply median filtering over these to restore the original image. This idea ensures fast filtering as only the noisy pixels are processed. Using simulations, we show that the proposed method scales efficiently with respect to accuracy and time. Through a combined measure of time and accuracy, we show that the proposed method exhibits nearly 42% improvement over the conventional ones.","PeriodicalId":322228,"journal":{"name":"2022 IEEE 5th International Conference on Image Processing Applications and Systems (IPAS)","volume":"Five 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130895020","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Visual Data Enciphering via DNA Encoding, S-Box, and Tent Mapping 通过DNA编码,S-Box和帐篷映射的可视化数据加密
Pub Date : 2022-12-05 DOI: 10.1109/IPAS55744.2022.10052832
Mohamed Gabr, H. Younis, Marwa Ibrahim, Sara Alajmy, Wassim Alexan
The ever-evolving nature of the Internet and wireless communications, as well as the production of huge amounts of multimedia every day has created a dire need for their security. In this paper, an image encryption technique that is based on 3 stages is proposed. The first stage makes use of DNA encoding. The second stage proposed and utilizes a novel S-box that is based on the Mersenne Twister and a linear descent algorithm. The third stage employs the Tent chaotic map. The computed performance evaluation metrics exhibit a high level of achieved security.
互联网和无线通信不断发展的本质,以及每天大量多媒体的产生,对它们的安全性产生了迫切的需求。本文提出了一种基于三阶段的图像加密技术。第一阶段利用DNA编码。第二阶段提出并利用了一种基于Mersenne Twister和线性下降算法的新型s盒。第三阶段使用帐篷混乱地图。计算的性能评估指标显示了高水平的已实现安全性。
{"title":"Visual Data Enciphering via DNA Encoding, S-Box, and Tent Mapping","authors":"Mohamed Gabr, H. Younis, Marwa Ibrahim, Sara Alajmy, Wassim Alexan","doi":"10.1109/IPAS55744.2022.10052832","DOIUrl":"https://doi.org/10.1109/IPAS55744.2022.10052832","url":null,"abstract":"The ever-evolving nature of the Internet and wireless communications, as well as the production of huge amounts of multimedia every day has created a dire need for their security. In this paper, an image encryption technique that is based on 3 stages is proposed. The first stage makes use of DNA encoding. The second stage proposed and utilizes a novel S-box that is based on the Mersenne Twister and a linear descent algorithm. The third stage employs the Tent chaotic map. The computed performance evaluation metrics exhibit a high level of achieved security.","PeriodicalId":322228,"journal":{"name":"2022 IEEE 5th International Conference on Image Processing Applications and Systems (IPAS)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133798838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Surface Crack Detection using Deep Convolutional Neural Network in Concrete Structures 基于深度卷积神经网络的混凝土结构表面裂纹检测
Pub Date : 2022-12-05 DOI: 10.1109/IPAS55744.2022.10052790
A. Rahai, M. Rahai, Mostafa Iraniparast, M. Ghatee
Regular safety inspections of concrete and steel structures during their serviceability are essential since they directly affect the reliability and structural health. Early detection of cracks helps prevent further damage. Traditional methods involve the detection of cracks by human visual inspection. However, it is difficult to visually find cracks and other defects for extremely large structures because of time and cost constraints. Therefore, the development of smart inspection systems has been given utmost importance. We provide a deep convolutional neural network (DCNN) with transfer learning (TF) technique for crack detection. To reduce false detection rates, the images used to train in the TF technique come from two different datasets (CCIC and SDNET). Moreover, the designed CNN is trained on 3200 images of $256 times 256$ pixel resolutions. Different deep learning networks are considered and the experiments on test images show that the accuracy of the damage detection is more than 99%. Results illustrate the viability of the suggested approach for crack observation and classification.
混凝土和钢结构在使用期间的定期安全检查是必要的,因为它们直接影响到结构的可靠性和健康。早期发现裂缝有助于防止进一步损坏。传统的方法是通过人的视觉检测来检测裂缝。然而,由于时间和成本的限制,很难在视觉上发现超大结构的裂缝和其他缺陷。因此,智能检测系统的发展被赋予了极大的重要性。提出了一种基于迁移学习(TF)技术的深度卷积神经网络(DCNN)用于裂纹检测。为了降低误检率,在TF技术中用于训练的图像来自两个不同的数据集(CCIC和SDNET)。此外,设计的CNN在3200张256 × 256像素分辨率的图像上进行训练。考虑了不同的深度学习网络,在测试图像上的实验表明,损伤检测的准确率在99%以上。结果表明,该方法对裂纹观测和分类是可行的。
{"title":"Surface Crack Detection using Deep Convolutional Neural Network in Concrete Structures","authors":"A. Rahai, M. Rahai, Mostafa Iraniparast, M. Ghatee","doi":"10.1109/IPAS55744.2022.10052790","DOIUrl":"https://doi.org/10.1109/IPAS55744.2022.10052790","url":null,"abstract":"Regular safety inspections of concrete and steel structures during their serviceability are essential since they directly affect the reliability and structural health. Early detection of cracks helps prevent further damage. Traditional methods involve the detection of cracks by human visual inspection. However, it is difficult to visually find cracks and other defects for extremely large structures because of time and cost constraints. Therefore, the development of smart inspection systems has been given utmost importance. We provide a deep convolutional neural network (DCNN) with transfer learning (TF) technique for crack detection. To reduce false detection rates, the images used to train in the TF technique come from two different datasets (CCIC and SDNET). Moreover, the designed CNN is trained on 3200 images of $256 times 256$ pixel resolutions. Different deep learning networks are considered and the experiments on test images show that the accuracy of the damage detection is more than 99%. Results illustrate the viability of the suggested approach for crack observation and classification.","PeriodicalId":322228,"journal":{"name":"2022 IEEE 5th International Conference on Image Processing Applications and Systems (IPAS)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125425204","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unrolling Alternating Direction Method of Multipliers for Visible and Infrared Image Fusion 可见与红外图像融合的乘法器交替方向展开方法
Pub Date : 2022-12-05 DOI: 10.1109/IPAS55744.2022.10052930
Altuğ Bakan, I. Erer
In this paper a new infrared and visible image fusion (IVIF) method which combines the advantages of optimization and deep learning based methods is proposed. This model takes the iterative solution used by the alternating direction method of the multiplier (ADMM) optimization method, and uses algorithm unrolling to obtain a high performance and efficient algorithm. Compared with traditional optimization methods, this model generates fusion with 99.6% improvement in terms of image fusion time, and compared with deep learning based algorithms, this model generates detailed fusion images with 99.1% improvement in terms of training time. Compared with the other state-of-the-art unrolling based methods, this model performs 26.7% better on average in terms of Average Gradient (AG), Cross Entropy (CE), Mutual Information (MI), Peak Signal-to-Noise Ratio (PSNR), and Structural Similarity Loss (SSIM) metrics with a minimal testing time cost.
本文提出了一种结合优化算法和深度学习算法优点的红外与可见光图像融合方法。该模型采用乘法器(ADMM)优化方法的交替方向法所采用的迭代解,并采用算法展开,得到了一种高性能、高效的算法。与传统优化方法相比,该模型生成的融合图像融合时间提高99.6%,与基于深度学习的算法相比,该模型生成的详细融合图像的训练时间提高99.1%。与其他最先进的基于展开的方法相比,该模型在平均梯度(AG)、交叉熵(CE)、互信息(MI)、峰值信噪比(PSNR)和结构相似度损失(SSIM)指标方面的平均性能提高了26.7%,且测试时间成本最小。
{"title":"Unrolling Alternating Direction Method of Multipliers for Visible and Infrared Image Fusion","authors":"Altuğ Bakan, I. Erer","doi":"10.1109/IPAS55744.2022.10052930","DOIUrl":"https://doi.org/10.1109/IPAS55744.2022.10052930","url":null,"abstract":"In this paper a new infrared and visible image fusion (IVIF) method which combines the advantages of optimization and deep learning based methods is proposed. This model takes the iterative solution used by the alternating direction method of the multiplier (ADMM) optimization method, and uses algorithm unrolling to obtain a high performance and efficient algorithm. Compared with traditional optimization methods, this model generates fusion with 99.6% improvement in terms of image fusion time, and compared with deep learning based algorithms, this model generates detailed fusion images with 99.1% improvement in terms of training time. Compared with the other state-of-the-art unrolling based methods, this model performs 26.7% better on average in terms of Average Gradient (AG), Cross Entropy (CE), Mutual Information (MI), Peak Signal-to-Noise Ratio (PSNR), and Structural Similarity Loss (SSIM) metrics with a minimal testing time cost.","PeriodicalId":322228,"journal":{"name":"2022 IEEE 5th International Conference on Image Processing Applications and Systems (IPAS)","volume":"15 3","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114048229","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cell tracking for live-cell microscopy using an activity-prioritized assignment strategy 使用活动优先分配策略的活细胞显微镜细胞跟踪
Pub Date : 2022-10-20 DOI: 10.1109/IPAS55744.2022.10053036
Karina Ruzaeva, J. Cohrs, Keitaro Kasahara, D. Kohlheyer, K. Nöh, B. Berkels
Cell tracking is an essential tool in live-cell imaging to determine single-cell features, such as division patterns or elongation rates. Unlike in common multiple object tracking, in microbial live-cell experiments cells are growing, moving, and dividing over time, to form cell colonies that are densely packed in mono-layer structures. With increasing cell numbers, following the precise cell-cell associations correctly over many generations becomes more and more challenging, due to the massively increasing number of possible associations. To tackle this challenge, we propose a fast parameter-free cell tracking approach, which consists of activity-prioritized nearest neighbor assignment of growing (expanding) cells and a combinatorial solver that assigns splitting mother cells to their daughters. As input for the tracking, Omnipose is utilized for instance segmentation. Unlike conventional nearest-neighbor-based tracking approaches, the assignment steps of our proposed method are based on a Gaussian activity-based metric, predicting the cell-specific migration probability, thereby limiting the number of erroneous assignments. In addition to being a building block for cell tracking, the proposed activity map is a standalone tracking-free metric for indicating cell activity. Finally, we perform a quantitative analysis of the tracking accuracy for different frame rates, to inform life scientists about a suitable (in terms of tracking performance) choice of the frame rate for their cultivation experiments, when cell tracks are the desired key outcome.
细胞跟踪是活细胞成像中确定单细胞特征(如分裂模式或延伸率)的重要工具。与常见的多目标跟踪不同,在微生物活细胞实验中,细胞随着时间的推移生长、移动和分裂,形成细胞菌落,这些菌落密集地排列在单层结构中。随着细胞数量的增加,由于可能的关联数量大量增加,在许多代中正确地跟踪精确的细胞-细胞关联变得越来越具有挑战性。为了解决这一挑战,我们提出了一种快速无参数细胞跟踪方法,该方法由生长(扩展)细胞的活性优先近邻分配和将分裂母细胞分配给其子细胞的组合求解器组成。作为跟踪的输入,Omnipose被用于实例分割。与传统的基于最近邻的跟踪方法不同,我们提出的方法的分配步骤是基于基于高斯活动的度量,预测细胞特异性迁移概率,从而限制错误分配的数量。除了作为细胞跟踪的构建块之外,所建议的活动映射还是一个独立的无跟踪度量,用于指示细胞活动。最后,我们对不同帧率的跟踪精度进行了定量分析,以告知生命科学家在他们的培养实验中选择合适的帧率(就跟踪性能而言),当细胞轨迹是期望的关键结果时。
{"title":"Cell tracking for live-cell microscopy using an activity-prioritized assignment strategy","authors":"Karina Ruzaeva, J. Cohrs, Keitaro Kasahara, D. Kohlheyer, K. Nöh, B. Berkels","doi":"10.1109/IPAS55744.2022.10053036","DOIUrl":"https://doi.org/10.1109/IPAS55744.2022.10053036","url":null,"abstract":"Cell tracking is an essential tool in live-cell imaging to determine single-cell features, such as division patterns or elongation rates. Unlike in common multiple object tracking, in microbial live-cell experiments cells are growing, moving, and dividing over time, to form cell colonies that are densely packed in mono-layer structures. With increasing cell numbers, following the precise cell-cell associations correctly over many generations becomes more and more challenging, due to the massively increasing number of possible associations. To tackle this challenge, we propose a fast parameter-free cell tracking approach, which consists of activity-prioritized nearest neighbor assignment of growing (expanding) cells and a combinatorial solver that assigns splitting mother cells to their daughters. As input for the tracking, Omnipose is utilized for instance segmentation. Unlike conventional nearest-neighbor-based tracking approaches, the assignment steps of our proposed method are based on a Gaussian activity-based metric, predicting the cell-specific migration probability, thereby limiting the number of erroneous assignments. In addition to being a building block for cell tracking, the proposed activity map is a standalone tracking-free metric for indicating cell activity. Finally, we perform a quantitative analysis of the tracking accuracy for different frame rates, to inform life scientists about a suitable (in terms of tracking performance) choice of the frame rate for their cultivation experiments, when cell tracks are the desired key outcome.","PeriodicalId":322228,"journal":{"name":"2022 IEEE 5th International Conference on Image Processing Applications and Systems (IPAS)","volume":"450 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123174704","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Likelihood ratio map for direct exoplanet detection 直接探测系外行星的似然比图
Pub Date : 2022-10-19 DOI: 10.1109/IPAS55744.2022.10052997
H. Daglayan, Simon Vary, F. Cantalloube, P. Absil, O. Absil
Direct imaging of exoplanets is a challenging task due to the small angular distance and high contrast relative to their host star, and the presence of quasi-static noise. We propose a new statistical method for direct imaging of exoplanets based on a likelihood ratio detection map, which assumes that the noise after the background subtraction step obeys a Laplacian distribution. We compare the method with two detection approaches based on signal-to-noise ratio (SNR) map after performing the background subtraction by the widely used Annular Principal Component Analysis (AnnPCA). The experimental results on the Beta Pictoris data set show the method outperforms SNR maps in terms of achieving the highest true positive rate (TPR) at zero false positive rate (FPR).
系外行星的直接成像是一项具有挑战性的任务,因为它们的角距很小,相对于其主星的对比度很高,而且存在准静态噪声。提出了一种基于似然比检测图的系外行星直接成像统计方法,该方法假设背景减除步骤后的噪声服从拉普拉斯分布。我们将该方法与两种基于信噪比(SNR)图的检测方法进行了比较,该方法采用了广泛使用的环形主成分分析(AnnPCA)进行背景减除。在Beta Pictoris数据集上的实验结果表明,该方法在零假阳性率(FPR)下实现最高真阳性率(TPR)方面优于信噪比图。
{"title":"Likelihood ratio map for direct exoplanet detection","authors":"H. Daglayan, Simon Vary, F. Cantalloube, P. Absil, O. Absil","doi":"10.1109/IPAS55744.2022.10052997","DOIUrl":"https://doi.org/10.1109/IPAS55744.2022.10052997","url":null,"abstract":"Direct imaging of exoplanets is a challenging task due to the small angular distance and high contrast relative to their host star, and the presence of quasi-static noise. We propose a new statistical method for direct imaging of exoplanets based on a likelihood ratio detection map, which assumes that the noise after the background subtraction step obeys a Laplacian distribution. We compare the method with two detection approaches based on signal-to-noise ratio (SNR) map after performing the background subtraction by the widely used Annular Principal Component Analysis (AnnPCA). The experimental results on the Beta Pictoris data set show the method outperforms SNR maps in terms of achieving the highest true positive rate (TPR) at zero false positive rate (FPR).","PeriodicalId":322228,"journal":{"name":"2022 IEEE 5th International Conference on Image Processing Applications and Systems (IPAS)","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114686104","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
2022 IEEE 5th International Conference on Image Processing Applications and Systems (IPAS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1