首页 > 最新文献

2020 International Conference on Machine Vision and Image Processing (MVIP)最新文献

英文 中文
A DNN-based Image Retrieval Approach for Detection of Defective Area in Carbon Fiber Reinforced Polymers through LDV Data 基于dnn的LDV数据检测碳纤维增强聚合物缺陷区域的图像检索方法
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116908
Erfan Basiri, Reza P. R. Hasanzadeh, Saman Hadi, M. Kersemans
Carbon fiber reinforced polymer (CFRP) materials, due to their specific strength and high consistency against erosion and corrosion, are widely used in industrial applications and high-tech engineering structures. However, there are also disadvantages: e.g. they are prone to different kinds of internal defects which could jeopardize the structural integrity of the CFRP material and therefore early detection of such defects can be an important task. Recently, local defect resonance (LDR), which is a subcategory of ultrasonic nondestructive testing, has been successfully used to solve this issue. However, the drawback of utilizing this technique is that the frequency at which the LDR occurs must be known. Further, the LDR-based technique has difficulty in assessing deep defects. In this paper, deep neural network (DNN) methodology is employed to remove this limitation and to acquire a better defect image retrieval process and also to achieve a model for the approximate depth estimation of such defects. In these regards, two types of defects called flat bottom holes (FBH) and barely visible impact damage (BVID) which are made in two CFRP coupons are used to evaluate the ability of the proposed method. Then, these two CFRPs are excited with a piezoelectric patch, and their corresponding laser Doppler vibrometry (LDV) response is collected through a scanning laser Doppler vibrometer (SLDV). Eventually, the superiority of our DNN-based approach is evaluated in comparison with other well-known classification methodologies.
碳纤维增强聚合物(CFRP)材料,由于其特定的强度和高一致性抗侵蚀和腐蚀,被广泛应用于工业应用和高科技工程结构。然而,也有缺点:例如,它们容易产生各种内部缺陷,这些缺陷可能危及CFRP材料的结构完整性,因此早期检测此类缺陷是一项重要任务。近年来,局部缺陷共振(LDR)作为超声无损检测的一个分支,成功地解决了这一问题。然而,使用这种技术的缺点是必须知道LDR发生的频率。此外,基于ldr的技术在评估深度缺陷方面存在困难。本文采用深度神经网络(deep neural network, DNN)方法消除了这一局限性,获得了更好的缺陷图像检索过程,并建立了缺陷深度近似估计模型。在这些方面,两种类型的缺陷称为平底孔(FBH)和几乎不可见的冲击损伤(BVID),这是在两个CFRP片材中产生的,用于评估所提出的方法的能力。然后,用压电贴片对这两个cfrp进行激励,并通过扫描式激光多普勒测振仪(SLDV)采集其相应的激光多普勒测振(LDV)响应。最后,与其他已知的分类方法进行比较,评估了基于dnn的方法的优越性。
{"title":"A DNN-based Image Retrieval Approach for Detection of Defective Area in Carbon Fiber Reinforced Polymers through LDV Data","authors":"Erfan Basiri, Reza P. R. Hasanzadeh, Saman Hadi, M. Kersemans","doi":"10.1109/MVIP49855.2020.9116908","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116908","url":null,"abstract":"Carbon fiber reinforced polymer (CFRP) materials, due to their specific strength and high consistency against erosion and corrosion, are widely used in industrial applications and high-tech engineering structures. However, there are also disadvantages: e.g. they are prone to different kinds of internal defects which could jeopardize the structural integrity of the CFRP material and therefore early detection of such defects can be an important task. Recently, local defect resonance (LDR), which is a subcategory of ultrasonic nondestructive testing, has been successfully used to solve this issue. However, the drawback of utilizing this technique is that the frequency at which the LDR occurs must be known. Further, the LDR-based technique has difficulty in assessing deep defects. In this paper, deep neural network (DNN) methodology is employed to remove this limitation and to acquire a better defect image retrieval process and also to achieve a model for the approximate depth estimation of such defects. In these regards, two types of defects called flat bottom holes (FBH) and barely visible impact damage (BVID) which are made in two CFRP coupons are used to evaluate the ability of the proposed method. Then, these two CFRPs are excited with a piezoelectric patch, and their corresponding laser Doppler vibrometry (LDV) response is collected through a scanning laser Doppler vibrometer (SLDV). Eventually, the superiority of our DNN-based approach is evaluated in comparison with other well-known classification methodologies.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134514756","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Low-cost 3D scanning using ultrasonic and camera data fusion for CNC Engraving Laser-Based Machine 利用超声波和相机数据融合的低成本3D扫描CNC激光雕刻机
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116903
M. J. Seikavandi
Sensor-fusion has gained much popularity in 3Dscanning in recent years. There are a variety of sensors like depth-camera, camera, laser-scanner, ultrasonic sensor, which are widely used in the area. The robotic research community has studied ultrasound sensors for decades. While they have lost attention with the advent of laser scanners and cameras, they remain successful for special applications due to their robustness and simplicity; additionally, ultrasound measurement is more robust than depth-camera for illumination-varying scenarios and work with glassy pieces. In this work, we choose camera and ultrasonic sensor fusion method for a CNC Engraving Machine concerning their lower cost and a particular act of applying. We use a heuristic, hand-crafted fusion to prepare 3D presentation of different pieces. The output data of the camera were down-sampled to ultrasonic data scale. By using an image processing method, the image, and ultrasonic data will be used to prepare a principle scheme and final 3D map.
近年来,传感器融合技术在三维扫描领域得到了广泛的应用。有各种各样的传感器,如深度相机,相机,激光扫描仪,超声波传感器,广泛应用于该领域。机器人研究界几十年来一直在研究超声波传感器。虽然随着激光扫描仪和相机的出现,它们已经失去了人们的关注,但由于它们的坚固性和简单性,它们在特殊应用中仍然取得了成功;此外,在光照变化的情况下,超声波测量比深度相机更可靠,并且可以处理玻璃片。在这项工作中,我们选择了相机和超声波传感器融合的方法在数控雕刻机上,因为它们的成本更低,并有一个特定的应用行为。我们使用启发式的、手工制作的融合来准备不同作品的3D呈现。相机的输出数据被下采样到超声波数据尺度。通过图像处理的方法,将图像和超声波数据结合起来,编制一个原理方案和最终的三维地图。
{"title":"Low-cost 3D scanning using ultrasonic and camera data fusion for CNC Engraving Laser-Based Machine","authors":"M. J. Seikavandi","doi":"10.1109/MVIP49855.2020.9116903","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116903","url":null,"abstract":"Sensor-fusion has gained much popularity in 3Dscanning in recent years. There are a variety of sensors like depth-camera, camera, laser-scanner, ultrasonic sensor, which are widely used in the area. The robotic research community has studied ultrasound sensors for decades. While they have lost attention with the advent of laser scanners and cameras, they remain successful for special applications due to their robustness and simplicity; additionally, ultrasound measurement is more robust than depth-camera for illumination-varying scenarios and work with glassy pieces. In this work, we choose camera and ultrasonic sensor fusion method for a CNC Engraving Machine concerning their lower cost and a particular act of applying. We use a heuristic, hand-crafted fusion to prepare 3D presentation of different pieces. The output data of the camera were down-sampled to ultrasonic data scale. By using an image processing method, the image, and ultrasonic data will be used to prepare a principle scheme and final 3D map.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"201 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116160440","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multiple-Vehicle Tracking in the Highway Using Appearance Model and Visual Object Tracking 基于外观模型和视觉目标跟踪的高速公路多车跟踪
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116905
Fateme Bafghi, B. Shoushtarian
In recent decades, due to the groundbreaking improvements in machine vision, many daily tasks are performed by computers. One of these tasks is multiple-vehicle tracking, which is widely used in different areas such as video surveillance and traffic monitoring. This paper focuses on introducing an efficient novel approach with acceptable accuracy. This is achieved through an efficient appearance and motion model based on the features extracted from each object. For this purpose, two different approaches have been used to extract features, i.e. features extracted from a deep neural network, and traditional features. Then the results from these two approaches are compared with state-of-the-art trackers. The results are obtained by executing the methods on the UA-DETRACK benchmark. The first method led to 58.9% accuracy while the second method caused up to 15.9%. The proposed methods can still be improved by extracting more distinguishable features.
近几十年来,由于机器视觉的突破性进步,许多日常任务都由计算机执行。其中一项任务是多车跟踪,它被广泛应用于视频监控和交通监控等不同领域。本文重点介绍了一种有效的、精度可接受的新方法。这是通过基于从每个对象中提取的特征的高效外观和运动模型来实现的。为此,我们采用了两种不同的方法来提取特征,即从深度神经网络中提取的特征和传统特征。然后将这两种方法的结果与最先进的跟踪器进行比较。结果通过在UA-DETRACK基准测试上执行这些方法得到。第一种方法的准确率为58.9%,第二种方法的准确率为15.9%。所提出的方法仍然可以通过提取更多可区分的特征来改进。
{"title":"Multiple-Vehicle Tracking in the Highway Using Appearance Model and Visual Object Tracking","authors":"Fateme Bafghi, B. Shoushtarian","doi":"10.1109/MVIP49855.2020.9116905","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116905","url":null,"abstract":"In recent decades, due to the groundbreaking improvements in machine vision, many daily tasks are performed by computers. One of these tasks is multiple-vehicle tracking, which is widely used in different areas such as video surveillance and traffic monitoring. This paper focuses on introducing an efficient novel approach with acceptable accuracy. This is achieved through an efficient appearance and motion model based on the features extracted from each object. For this purpose, two different approaches have been used to extract features, i.e. features extracted from a deep neural network, and traditional features. Then the results from these two approaches are compared with state-of-the-art trackers. The results are obtained by executing the methods on the UA-DETRACK benchmark. The first method led to 58.9% accuracy while the second method caused up to 15.9%. The proposed methods can still be improved by extracting more distinguishable features.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127394040","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Exploring the Gradient for Video Quality Assessment 视频质量评价的梯度研究
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116869
Hossein Motamednia, Pooryaa Cheraaqee, Azadeh Mansouri
This paper presents an algorithm which incorporates spatial and temporal gradients for full reference video quality assessment. In the proposed method the frame-based gradient magnitude similarity deviation is calculated to form the spatial quality vector. To capture the temporal distortion, the similarity of frame difference is measured. In the proposed method, we extract the worst scores in both the spatial and temporal vectors by introducing the variable-length temporal window for max-pooling operation. The resultant vectors are then combined to form the final score. The performance of the proposed method is evaluated on LIVE SD and EPFL- PoliMI datasets. The results clearly illustrate that, despite the computational efficiency, the predictions are highly correlated with human visual system.
本文提出了一种结合时空梯度的全参考视频质量评估算法。该方法计算基于帧的梯度震级相似偏差,形成空间质量矢量。为了捕获时间畸变,测量了帧差的相似度。在该方法中,我们通过引入变长时间窗口进行最大池化操作,提取空间和时间向量上的最差分数。然后将结果向量组合起来形成最终分数。在LIVE SD和EPFL- PoliMI数据集上对该方法的性能进行了评价。结果清楚地表明,尽管计算效率高,但预测与人类视觉系统高度相关。
{"title":"Exploring the Gradient for Video Quality Assessment","authors":"Hossein Motamednia, Pooryaa Cheraaqee, Azadeh Mansouri","doi":"10.1109/MVIP49855.2020.9116869","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116869","url":null,"abstract":"This paper presents an algorithm which incorporates spatial and temporal gradients for full reference video quality assessment. In the proposed method the frame-based gradient magnitude similarity deviation is calculated to form the spatial quality vector. To capture the temporal distortion, the similarity of frame difference is measured. In the proposed method, we extract the worst scores in both the spatial and temporal vectors by introducing the variable-length temporal window for max-pooling operation. The resultant vectors are then combined to form the final score. The performance of the proposed method is evaluated on LIVE SD and EPFL- PoliMI datasets. The results clearly illustrate that, despite the computational efficiency, the predictions are highly correlated with human visual system.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130128600","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Quality Assessment for Retargeted Images: A Review 重定向图像的质量评估:综述
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116899
Maryam Karimi, Erfan Entezami
Transmission, saving and many processing methods cause different damage in images. Image Quality Assessment (IQA) is necessary to benchmark processing algorithms, to optimize them, and to monitor the quality of images in quality control systems. Traditional quality metrics have low correlations with subjective perception. The key problem is to evaluate the distorted images as human do. Subjective quality assessment is more reliable but is cumbersome and time-consuming, so it is impossible to embed it in online applications. Therefore, many objective perceptual IQA models have been developed until now. Content-aware retargeting methods aim to adapt source images to target display devices with different sizes and aspect ratios so that salient areas will be less distorted. The size mismatch and the completely different distortions caused by retargeting have made common IQA methods useless in this area. Therefore, retargeted Image Quality Assessment (RIQA) methods are designed for this purpose. The quality of retargeted images is different depending to image content and retargeting algorithm. This paper provides a literature review and a new categorization of the current subjective and objective retargeted image quality measures. Also, we intend to compare and analyze the performance of these measures. It is demonstrated that the performance of RIQA methods can be further improved by using high-level descriptors in addition to low-level ones.
传输、保存和许多处理方法都会对图像造成不同程度的损害。在质量控制系统中,图像质量评估(IQA)是对处理算法进行基准测试、优化和监控图像质量的必要手段。传统的质量指标与主观感知的相关性较低。关键问题是如何像人一样对失真图像进行评估。主观质量评估更可靠,但繁琐且耗时,因此无法将其嵌入在线应用程序。因此,到目前为止,已经开发了许多客观的感知IQA模型。内容感知重定向方法旨在使源图像适应具有不同尺寸和宽高比的目标显示设备,从而减少显著区域的失真。由于重定向引起的大小不匹配和完全不同的扭曲,使得常见的IQA方法在这一领域毫无用处。因此,重定向图像质量评估(RIQA)方法是为此目的而设计的。根据图像内容和重定向算法的不同,重定向图像的质量也不同。本文对现有的主观和客观重定向图像质量测量方法进行了综述和分类。此外,我们打算比较和分析这些措施的表现。结果表明,除了使用低级描述符外,还使用高级描述符可以进一步提高RIQA方法的性能。
{"title":"Quality Assessment for Retargeted Images: A Review","authors":"Maryam Karimi, Erfan Entezami","doi":"10.1109/MVIP49855.2020.9116899","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116899","url":null,"abstract":"Transmission, saving and many processing methods cause different damage in images. Image Quality Assessment (IQA) is necessary to benchmark processing algorithms, to optimize them, and to monitor the quality of images in quality control systems. Traditional quality metrics have low correlations with subjective perception. The key problem is to evaluate the distorted images as human do. Subjective quality assessment is more reliable but is cumbersome and time-consuming, so it is impossible to embed it in online applications. Therefore, many objective perceptual IQA models have been developed until now. Content-aware retargeting methods aim to adapt source images to target display devices with different sizes and aspect ratios so that salient areas will be less distorted. The size mismatch and the completely different distortions caused by retargeting have made common IQA methods useless in this area. Therefore, retargeted Image Quality Assessment (RIQA) methods are designed for this purpose. The quality of retargeted images is different depending to image content and retargeting algorithm. This paper provides a literature review and a new categorization of the current subjective and objective retargeted image quality measures. Also, we intend to compare and analyze the performance of these measures. It is demonstrated that the performance of RIQA methods can be further improved by using high-level descriptors in addition to low-level ones.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"138 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122624111","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
MVIP 2020 Table of Contents MVIP 2020目录
Pub Date : 2020-02-01 DOI: 10.1109/mvip49855.2020.9116904
{"title":"MVIP 2020 Table of Contents","authors":"","doi":"10.1109/mvip49855.2020.9116904","DOIUrl":"https://doi.org/10.1109/mvip49855.2020.9116904","url":null,"abstract":"","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"2008 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123916642","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Attention-Based Face AntiSpoofing of RGB Camera using a Minimal End-2-End Neural Network 基于注意力的RGB相机人脸防欺骗最小端到端神经网络
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116872
A. Ghofrani, Rahil Mahdian Toroghi, Seyed Mojtaba Tabatabaie
Face anti-spoofing aims at identifying the real face, as well as the fake one, and gains a high attention in security sensitive applications, liveness detection, fingerprinting, and so on. In this paper, we address the anti-spoofing problem by proposing two end-to-end systems of convolutional neural networks. One model is developed based on the EfficientNet B0 network which has been modified in the final dense layers. The second one, is a very light model of the MobileNet V2, which has been contracted, modified and retrained efficiently on the data being created based on the Rose-Youtu dataset, for this purpose. The experiments show that, both of the proposed architectures achieve remarkable results on detecting the real and fake images of the face input data. The experiments clearly show that the heavy-weight model could be efficiently employed in server side implementations, whereas the low-weight model could be easily implemented on the hand-held devices and both perform perfectly well using merely RGB input images.
人脸防欺骗技术旨在识别人脸的真伪,在安全敏感应用、活体检测、指纹识别等领域受到高度关注。在本文中,我们通过提出两个卷积神经网络的端到端系统来解决反欺骗问题。其中一个模型是基于在最终密集层中进行了修改的EfficientNet B0网络开发的。第二个,是MobileNet V2的一个非常轻的模型,为了这个目的,它已经在基于Rose-Youtu数据集创建的数据上进行了有效的收缩、修改和再训练。实验表明,这两种架构在人脸输入数据的真假图像检测上都取得了显著的效果。实验清楚地表明,重权重模型可以有效地用于服务器端实现,而低权重模型可以很容易地在手持设备上实现,并且仅使用RGB输入图像就可以很好地执行。
{"title":"Attention-Based Face AntiSpoofing of RGB Camera using a Minimal End-2-End Neural Network","authors":"A. Ghofrani, Rahil Mahdian Toroghi, Seyed Mojtaba Tabatabaie","doi":"10.1109/MVIP49855.2020.9116872","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116872","url":null,"abstract":"Face anti-spoofing aims at identifying the real face, as well as the fake one, and gains a high attention in security sensitive applications, liveness detection, fingerprinting, and so on. In this paper, we address the anti-spoofing problem by proposing two end-to-end systems of convolutional neural networks. One model is developed based on the EfficientNet B0 network which has been modified in the final dense layers. The second one, is a very light model of the MobileNet V2, which has been contracted, modified and retrained efficiently on the data being created based on the Rose-Youtu dataset, for this purpose. The experiments show that, both of the proposed architectures achieve remarkable results on detecting the real and fake images of the face input data. The experiments clearly show that the heavy-weight model could be efficiently employed in server side implementations, whereas the low-weight model could be easily implemented on the hand-held devices and both perform perfectly well using merely RGB input images.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"138 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123727142","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A weighted, statistical based, No-Reference metric for holography Image Quality Assessment 一种加权的、基于统计的、无参考的全息图像质量评价方法
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116927
Vahid Hajihashemi, Mohammad Mehdi Arab Ameri, A. Alavi Gharahbagh, Hassan Pahlouvary
Digital holography is one of the 3D imaging systems that suffer Speckle noise. With respect to the importance of quality in 3D images, we develop an efficient general-purpose blind/no-reference holography image quality assessment metric for evaluating the quality of digital holography images. The main novelty of our approach to blind image quality assessment is based on the hypothesis that each digital holography has statistical properties that are changing in the presence of speckle noise. This change can be measured by some full reference metrics that are applied to input image and a new image, which were made by adding a known level of speckle noise to input image. These full reference measurements have the ability of identifying the distortion afflicting the input image and perform a no-reference quality assessment. In fact, adding noise to input image leads to quality loss, and the value of this loss give information about the input image quality. Finally, the result of the proposed method in estimating the quality of digital holography images were compared with some well-known full reference methods in order to demonstrate its ability.
数字全息是一种受散斑噪声影响的3D成像系统。考虑到三维图像质量的重要性,我们开发了一种有效的通用盲/无参考全息图像质量评估指标,用于评估数字全息图像的质量。我们盲图像质量评估方法的主要新颖之处是基于这样一个假设,即每个数字全息图都具有在散斑噪声存在下发生变化的统计特性。这种变化可以通过应用于输入图像和新图像的一些完整参考指标来测量,这些指标是通过向输入图像添加已知水平的散斑噪声而产生的。这些全参考测量具有识别影响输入图像的失真和执行无参考质量评估的能力。实际上,在输入图像中加入噪声会导致质量损失,而这种损失的值给出了输入图像质量的信息。最后,将所提方法用于数字全息图像质量估计的结果与一些已知的全参考方法进行了比较,以证明所提方法的能力。
{"title":"A weighted, statistical based, No-Reference metric for holography Image Quality Assessment","authors":"Vahid Hajihashemi, Mohammad Mehdi Arab Ameri, A. Alavi Gharahbagh, Hassan Pahlouvary","doi":"10.1109/MVIP49855.2020.9116927","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116927","url":null,"abstract":"Digital holography is one of the 3D imaging systems that suffer Speckle noise. With respect to the importance of quality in 3D images, we develop an efficient general-purpose blind/no-reference holography image quality assessment metric for evaluating the quality of digital holography images. The main novelty of our approach to blind image quality assessment is based on the hypothesis that each digital holography has statistical properties that are changing in the presence of speckle noise. This change can be measured by some full reference metrics that are applied to input image and a new image, which were made by adding a known level of speckle noise to input image. These full reference measurements have the ability of identifying the distortion afflicting the input image and perform a no-reference quality assessment. In fact, adding noise to input image leads to quality loss, and the value of this loss give information about the input image quality. Finally, the result of the proposed method in estimating the quality of digital holography images were compared with some well-known full reference methods in order to demonstrate its ability.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114666306","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Automatic Skin Cancer (Melanoma) Detection by Processing Dermatoscopic images 通过处理皮肤镜图像自动检测皮肤癌(黑色素瘤)
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116918
Hadi Moazen, M. Jamzad
Melanoma is the deadliest form of skin cancer if not treated early. The best way to cure melanoma is to treat it in its earliest stage of development. Since melanoma is similar to benign moles in its shape and appearance, it is often mistaken for moles and left untreated. Automatic melanoma detection is an essential way to increase the survival rate of patients by detecting melanoma in its early stages. In this paper, a new method for automatic diagnosis of melanoma using segmented dermatoscopic images is provided. Almost all related methods follow similar approaches but using different features. We have introduced several new features which could improve the accuracy of diagnosing melanoma. For evaluation we have implemented and tested all methods on the ISIC archive, which is the largest openly available dataset of dermatoscopic melanoma images. Our method outperforms most recent previous works’ accuracy on the ISIC dataset by 1.5 percent. It also achieves a 2.32-point higher F1 score while obtaining a comparable sensitivity.
如果不及早治疗,黑色素瘤是最致命的皮肤癌。治疗黑色素瘤的最好方法是在其发展的早期阶段进行治疗。由于黑色素瘤在形状和外观上与良性痣相似,因此经常被误认为是痣而不进行治疗。黑色素瘤的自动检测是提高患者生存率的重要途径,可以在黑色素瘤的早期阶段进行检测。本文提出了一种基于皮肤镜图像分割的黑色素瘤自动诊断方法。几乎所有相关的方法都遵循类似的方法,但使用了不同的功能。我们引入了一些新的特征,可以提高诊断黑色素瘤的准确性。为了评估,我们在ISIC档案上实施并测试了所有方法,这是最大的公开可用的皮肤镜下黑色素瘤图像数据集。我们的方法比以前在ISIC数据集上的最新作品的准确性高出1.5%。它还实现了2.32分更高的F1得分,同时获得相当的灵敏度。
{"title":"Automatic Skin Cancer (Melanoma) Detection by Processing Dermatoscopic images","authors":"Hadi Moazen, M. Jamzad","doi":"10.1109/MVIP49855.2020.9116918","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116918","url":null,"abstract":"Melanoma is the deadliest form of skin cancer if not treated early. The best way to cure melanoma is to treat it in its earliest stage of development. Since melanoma is similar to benign moles in its shape and appearance, it is often mistaken for moles and left untreated. Automatic melanoma detection is an essential way to increase the survival rate of patients by detecting melanoma in its early stages. In this paper, a new method for automatic diagnosis of melanoma using segmented dermatoscopic images is provided. Almost all related methods follow similar approaches but using different features. We have introduced several new features which could improve the accuracy of diagnosing melanoma. For evaluation we have implemented and tested all methods on the ISIC archive, which is the largest openly available dataset of dermatoscopic melanoma images. Our method outperforms most recent previous works’ accuracy on the ISIC dataset by 1.5 percent. It also achieves a 2.32-point higher F1 score while obtaining a comparable sensitivity.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121435495","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Class Attention Map Distillation for Efficient Semantic Segmentation 类注意图蒸馏的高效语义分割
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116875
Nader Karimi Bavandpour, S. Kasaei
In this paper, a novel method for capturing the information of a powerful and trained deep convolutional neural network and distilling it into a training smaller network is proposed. This is the first time that a saliency map method is employed to extract useful knowledge from a convolutional neural network for distillation. This method, despite of many others which work on final layers, can successfully extract suitable information for distillation from intermediate layers of a network by making class specific attention maps and then forcing the student network to mimic producing those attentions. This novel knowledge distillation training is implemented using state-of-the-art DeepLab and PSPNet segmentation networks and its effectiveness is shown by experiments on the standard Pascal Voc 2012 dataset.
本文提出了一种新的方法,用于捕获强大且经过训练的深度卷积神经网络的信息并将其提取到训练较小的网络中。这是首次采用显著性映射方法从卷积神经网络中提取有用的知识进行提炼。尽管有许多其他方法在最后一层工作,但这种方法可以通过制作班级特定的注意力图,然后迫使学生网络模仿产生这些注意力,成功地从网络的中间层提取合适的信息进行蒸馏。利用最先进的DeepLab和PSPNet分割网络实现了这种新颖的知识蒸馏训练,并通过在标准Pascal Voc 2012数据集上的实验证明了其有效性。
{"title":"Class Attention Map Distillation for Efficient Semantic Segmentation","authors":"Nader Karimi Bavandpour, S. Kasaei","doi":"10.1109/MVIP49855.2020.9116875","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116875","url":null,"abstract":"In this paper, a novel method for capturing the information of a powerful and trained deep convolutional neural network and distilling it into a training smaller network is proposed. This is the first time that a saliency map method is employed to extract useful knowledge from a convolutional neural network for distillation. This method, despite of many others which work on final layers, can successfully extract suitable information for distillation from intermediate layers of a network by making class specific attention maps and then forcing the student network to mimic producing those attentions. This novel knowledge distillation training is implemented using state-of-the-art DeepLab and PSPNet segmentation networks and its effectiveness is shown by experiments on the standard Pascal Voc 2012 dataset.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124145571","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2020 International Conference on Machine Vision and Image Processing (MVIP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1