首页 > 最新文献

Signal and image processing : an international journal最新文献

英文 中文
Target Detection and Classification Improvements using Contrast Enhanced 16-bit Infrared Videos 使用对比度增强的16位红外视频改进目标检测和分类
Pub Date : 2021-02-28 DOI: 10.5121/SIPIJ.2021.12103
C. Kwan, David Gribben
In our earlier target detection and classification papers, we used 8-bit infrared videos in the Defense Systems Information Analysis Center(DSIAC) video dataset. In this paper, we focus on how we can improve the target detection and classification results using 16-bit videos. One problem with the 16-bit videos is that some image frames have very low contrast. Two methods were explored to improve upon previous detection and classification results. The first method used to improve contrast was effectively the same as the baseline 8-bit video data but using the 16-bit raw data rather than the 8-bit data taken from the avi files. The second method used was a second order histogram matching algorithm that preserves the 16-bit nature of the videos while providing normalization and contrast enhancement. Results showed the second order histogram matching algorithm improved the target detection using You Only Look Once (YOLO) and classificationusing Residual Network (ResNet) performance. The average precision (AP) metric in YOLO was improved by 8%. This is quite significant. The overall accuracy (OA) of ResNet has been improved by 12%. This is also very significant.
在我们早期的目标检测和分类论文中,我们使用了国防系统信息分析中心(DSIAC)视频数据集中的8位红外视频。在本文中,我们重点研究了如何使用16位视频来改进目标检测和分类结果。16位视频的一个问题是一些图像帧的对比度很低。我们探索了两种方法来改进之前的检测和分类结果。用于提高对比度的第一种方法有效地与基线8位视频数据相同,但使用16位原始数据而不是从avi文件中获取的8位数据。第二种方法是二阶直方图匹配算法,它保留了视频的16位特性,同时提供了归一化和对比度增强。结果表明,二阶直方图匹配算法提高了使用You Only Look Once (YOLO)的目标检测性能和使用Residual Network (ResNet)的分类性能。YOLO的平均精度(AP)指标提高了8%。这是非常重要的。ResNet的整体准确率(OA)提高了12%。这也是非常重要的。
{"title":"Target Detection and Classification Improvements using Contrast Enhanced 16-bit Infrared Videos","authors":"C. Kwan, David Gribben","doi":"10.5121/SIPIJ.2021.12103","DOIUrl":"https://doi.org/10.5121/SIPIJ.2021.12103","url":null,"abstract":"In our earlier target detection and classification papers, we used 8-bit infrared videos in the Defense Systems Information Analysis Center(DSIAC) video dataset. In this paper, we focus on how we can improve the target detection and classification results using 16-bit videos. One problem with the 16-bit videos is that some image frames have very low contrast. Two methods were explored to improve upon previous detection and classification results. The first method used to improve contrast was effectively the same as the baseline 8-bit video data but using the 16-bit raw data rather than the 8-bit data taken from the avi files. The second method used was a second order histogram matching algorithm that preserves the 16-bit nature of the videos while providing normalization and contrast enhancement. Results showed the second order histogram matching algorithm improved the target detection using You Only Look Once (YOLO) and classificationusing Residual Network (ResNet) performance. The average precision (AP) metric in YOLO was improved by 8%. This is quite significant. The overall accuracy (OA) of ResNet has been improved by 12%. This is also very significant.","PeriodicalId":90726,"journal":{"name":"Signal and image processing : an international journal","volume":"108 1","pages":"23-38"},"PeriodicalIF":0.0,"publicationDate":"2021-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85677560","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Modelling, Conception and Simulation of a Digital Watermarking System based on Hyperbolic Geometry 基于双曲几何的数字水印系统建模、构想与仿真
Pub Date : 2021-01-01 DOI: 10.5121/sipij.2021.12401
Coulibaly Cheick Yacouba Rachid, Tiendrebeogo B. Telesphore
The digital revolution has increased the production and exchange of high-value documents between institutions, businesses and the general public. In order to secure these exchanges, it is essential to guarantee the authenticity, integrity and ownership of these documents. Digital watermarking is a possible solution to this challenge as it has already been used for copyright protection, source tracking and video authentication. It also provides integrity protection, which is useful for many types of documents (official documents, medical images). In this paper, we propose a new watermarking solution applicable to images and based on the hyperbolic geometry. Our new solution is based on existing work in the field of digital watermarking
数字革命增加了机构、企业和公众之间高价值文件的生产和交换。为了确保这些交换的安全,必须保证这些文件的真实性、完整性和所有权。数字水印是应对这一挑战的一种可能的解决方案,因为它已经被用于版权保护、源跟踪和视频认证。它还提供完整性保护,这对许多类型的文件(官方文件、医学图像)都很有用。本文提出了一种新的基于双曲几何的图像水印方案。我们的新方案是基于现有的数字水印领域的工作
{"title":"Modelling, Conception and Simulation of a Digital Watermarking System based on Hyperbolic Geometry","authors":"Coulibaly Cheick Yacouba Rachid, Tiendrebeogo B. Telesphore","doi":"10.5121/sipij.2021.12401","DOIUrl":"https://doi.org/10.5121/sipij.2021.12401","url":null,"abstract":"The digital revolution has increased the production and exchange of high-value documents between institutions, businesses and the general public. In order to secure these exchanges, it is essential to guarantee the authenticity, integrity and ownership of these documents. Digital watermarking is a possible solution to this challenge as it has already been used for copyright protection, source tracking and video authentication. It also provides integrity protection, which is useful for many types of documents (official documents, medical images). In this paper, we propose a new watermarking solution applicable to images and based on the hyperbolic geometry. Our new solution is based on existing work in the field of digital watermarking","PeriodicalId":90726,"journal":{"name":"Signal and image processing : an international journal","volume":"84 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83864472","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Novel Graph Representation for Skeleton-based Action Recognition 一种新的基于骨架的动作识别图表示方法
Pub Date : 2020-12-30 DOI: 10.5121/SIPIJ.2020.11605
Tingwei Li, Ruiwen Zhang, Qing Li
Graph convolutional networks (GCNs) have been proven to be effective for processing structured data, so that it can effectively capture the features of related nodes and improve the performance of model. More attention is paid to employing GCN in Skeleton-Based action recognition. But there are some challenges with the existing methods based on GCNs. First, the consistency of temporal and spatial features is ignored due to extracting features node by node and frame by frame. We design a generic representation of skeleton sequences for action recognition and propose a novel model called Temporal Graph Networks (TGN), which can obtain spatiotemporal features simultaneously. Secondly, the adjacency matrix of graph describing the relation of joints are mostly depended on the physical connection between joints. We propose a multi-scale graph strategy to appropriately describe the relations between joints in skeleton graph, which adopts a full-scale graph, part-scale graph and core-scale graph to capture the local features of each joint and the contour features of important joints. Extensive experiments are conducted on two large datasets including NTU RGB+D and Kinetics Skeleton. And the experiments results show that TGN with our graph strategy outperforms other state-of-the-art methods.
图卷积网络(Graph convolutional networks, GCNs)在结构化数据的处理上已经被证明是有效的,它可以有效地捕捉相关节点的特征,提高模型的性能。将GCN应用于基于骨骼的动作识别得到了更多的关注。但是现有的基于GCNs的方法存在一些挑战。首先,采用逐节点、逐帧提取特征的方法,忽略了时空特征的一致性;我们设计了一种用于动作识别的骨架序列的通用表示,并提出了一种可以同时获取时空特征的新模型——时间图网络(TGN)。其次,描述节点关系的图的邻接矩阵大多依赖于节点之间的物理连接。为了在骨架图中恰当地描述关节之间的关系,我们提出了一种多尺度图策略,该策略采用全尺度图、部分尺度图和核心尺度图来捕捉每个关节的局部特征和重要关节的轮廓特征。在NTU RGB+D和Kinetics Skeleton两个大型数据集上进行了广泛的实验。实验结果表明,基于我们的图策略的TGN优于其他最先进的方法。
{"title":"A Novel Graph Representation for Skeleton-based Action Recognition","authors":"Tingwei Li, Ruiwen Zhang, Qing Li","doi":"10.5121/SIPIJ.2020.11605","DOIUrl":"https://doi.org/10.5121/SIPIJ.2020.11605","url":null,"abstract":"Graph convolutional networks (GCNs) have been proven to be effective for processing structured data, so that it can effectively capture the features of related nodes and improve the performance of model. More attention is paid to employing GCN in Skeleton-Based action recognition. But there are some challenges with the existing methods based on GCNs. First, the consistency of temporal and spatial features is ignored due to extracting features node by node and frame by frame. We design a generic representation of skeleton sequences for action recognition and propose a novel model called Temporal Graph Networks (TGN), which can obtain spatiotemporal features simultaneously. Secondly, the adjacency matrix of graph describing the relation of joints are mostly depended on the physical connection between joints. We propose a multi-scale graph strategy to appropriately describe the relations between joints in skeleton graph, which adopts a full-scale graph, part-scale graph and core-scale graph to capture the local features of each joint and the contour features of important joints. Extensive experiments are conducted on two large datasets including NTU RGB+D and Kinetics Skeleton. And the experiments results show that TGN with our graph strategy outperforms other state-of-the-art methods.","PeriodicalId":90726,"journal":{"name":"Signal and image processing : an international journal","volume":"8 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75514286","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Facial Age Estimation using Transfer Learning and Bayesian Optimization based on Gender Information 基于性别信息的迁移学习和贝叶斯优化面部年龄估计
Pub Date : 2020-12-30 DOI: 10.5121/SIPIJ.2020.11604
Marwa Ahmed, Serestina Viriri
Age estimation of unrestricted imaging circumstances has attracted an augmented recognition as it is appropriate in several real-world applications such as surveillance, face recognition, age synthesis, access control, and electronic customer relationship management. Current deep learning-based methods have displayed encouraging performance in age estimation field. Males and Females have a variable type of appearance aging pattern; this results in age differently. This fact leads to assuming that using gender information may improve the age estimator performance. We have proposed a novel model based on Gender Classification. A Convolutional Neural Network (CNN) is used to get Gender Information, then Bayesian Optimization is applied to this pre-trained CNN when fine-tuned for age estimation task. Bayesian Optimization reduces the classification error on the validation set for the pre-trained model. Extensive experiments are done to assess our proposed model on two data sets: FERET and FG-NET. The experiments’ result indicates that using a pre-trained CNN containing Gender Information with Bayesian Optimization outperforms the state of the arts on FERET and FG-NET data sets with a Mean Absolute Error (MAE) of 1.2 and 2.67 respectively.
不受限制的成像环境的年龄估计已经吸引了增强识别,因为它适用于几个现实世界的应用,如监视、人脸识别、年龄合成、访问控制和电子客户关系管理。目前基于深度学习的方法在年龄估计领域显示出令人鼓舞的表现。男性和女性具有可变类型的外观老化模式;这导致了年龄的不同。这一事实导致假设使用性别信息可能会提高年龄估计器的性能。我们提出了一个基于性别分类的新模型。首先使用卷积神经网络(CNN)获取性别信息,然后对预训练好的CNN进行微调,进行年龄估计任务的贝叶斯优化。贝叶斯优化减少了预训练模型在验证集上的分类误差。在两个数据集:FERET和FG-NET上进行了大量的实验来评估我们提出的模型。实验结果表明,使用贝叶斯优化的包含性别信息的预训练CNN在FERET和FG-NET数据集上的平均绝对误差(MAE)分别为1.2和2.67,优于目前的研究水平。
{"title":"Facial Age Estimation using Transfer Learning and Bayesian Optimization based on Gender Information","authors":"Marwa Ahmed, Serestina Viriri","doi":"10.5121/SIPIJ.2020.11604","DOIUrl":"https://doi.org/10.5121/SIPIJ.2020.11604","url":null,"abstract":"Age estimation of unrestricted imaging circumstances has attracted an augmented recognition as it is appropriate in several real-world applications such as surveillance, face recognition, age synthesis, access control, and electronic customer relationship management. Current deep learning-based methods have displayed encouraging performance in age estimation field. Males and Females have a variable type of appearance aging pattern; this results in age differently. This fact leads to assuming that using gender information may improve the age estimator performance. We have proposed a novel model based on Gender Classification. A Convolutional Neural Network (CNN) is used to get Gender Information, then Bayesian Optimization is applied to this pre-trained CNN when fine-tuned for age estimation task. Bayesian Optimization reduces the classification error on the validation set for the pre-trained model. Extensive experiments are done to assess our proposed model on two data sets: FERET and FG-NET. The experiments’ result indicates that using a pre-trained CNN containing Gender Information with Bayesian Optimization outperforms the state of the arts on FERET and FG-NET data sets with a Mean Absolute Error (MAE) of 1.2 and 2.67 respectively.","PeriodicalId":90726,"journal":{"name":"Signal and image processing : an international journal","volume":"113 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82437669","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Neighbour Local Variability for Multi-Focus Images Fusion 多焦点图像融合的邻居局部变异
Pub Date : 2020-12-30 DOI: 10.5121/SIPIJ.2020.11603
I. Wahyuni, R. Sabre
The goal of multi-focus image fusion is to integrate images with different focus objects in order to obtain a single image with all focus objects. In this paper, we give a new method based on neighbour local variability (NLV) to fuse multi-focus images. At each pixel, the method uses the local variability calculated from the quadratic difference between the value of the pixel and the value of all pixels in its neighbourhood. It expresses the behaviour of the pixel with respect to its neighbours. The variability preserves the edge function because it detects the sharp intensity of the image. The proposed fusion of each pixel consists of weighting each pixel by the exponential of its local variability. The quality of this fusion depends on the size of the neighbourhood region considered. The size depends on the variance and the size of the blur filter. We start by modelling the value of the neighbourhood region size as a function of the variance and the size of the blur filter. We compare our method to other methods given in the literature. We show that our method gives a better result.
多焦点图像融合的目标是将具有不同焦点对象的图像进行融合,以获得具有所有焦点对象的单一图像。本文提出了一种基于邻域局部变异(NLV)的多焦点图像融合方法。在每个像素上,该方法使用由像素值与其邻域所有像素值之间的二次差计算的局部变异性。它表示像素相对于其邻居的行为。可变性保留了边缘函数,因为它检测图像的尖锐强度。所提出的每个像素的融合包括对每个像素的局部变异指数进行加权。这种融合的质量取决于所考虑的邻近区域的大小。大小取决于方差和模糊滤镜的大小。我们首先将邻域大小的值建模为方差和模糊过滤器大小的函数。我们将我们的方法与文献中给出的其他方法进行比较。结果表明,该方法能得到较好的结果。
{"title":"Neighbour Local Variability for Multi-Focus Images Fusion","authors":"I. Wahyuni, R. Sabre","doi":"10.5121/SIPIJ.2020.11603","DOIUrl":"https://doi.org/10.5121/SIPIJ.2020.11603","url":null,"abstract":"The goal of multi-focus image fusion is to integrate images with different focus objects in order to obtain a single image with all focus objects. In this paper, we give a new method based on neighbour local variability (NLV) to fuse multi-focus images. At each pixel, the method uses the local variability calculated from the quadratic difference between the value of the pixel and the value of all pixels in its neighbourhood. It expresses the behaviour of the pixel with respect to its neighbours. The variability preserves the edge function because it detects the sharp intensity of the image. The proposed fusion of each pixel consists of weighting each pixel by the exponential of its local variability. The quality of this fusion depends on the size of the neighbourhood region considered. The size depends on the variance and the size of the blur filter. We start by modelling the value of the neighbourhood region size as a function of the variance and the size of the blur filter. We compare our method to other methods given in the literature. We show that our method gives a better result.","PeriodicalId":90726,"journal":{"name":"Signal and image processing : an international journal","volume":"16 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79246865","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Further Improvements of CFA 3.0 by Combining Inpainting and Pansharpening Techniques 结合Inpainting和Pansharpening技术对CFA 3.0的进一步改进
Pub Date : 2020-12-30 DOI: 10.5121/SIPIJ.2020.11601
C. Kwan, Jude Larkin
Color Filter Array (CFA) has been widely used in digital cameras. There are many variants of CFAs in the literature. Recently, a new CFA known as CFA 3.0 was proposed by us and has been shown to yield reasonable performance as compared to some standard ones. In this paper, we investigate the use of inpainting algorithms to further improve the demosaicing performance of CFA 3.0. Six conventional and deep learning based inpainting algorithms were compared. Extensive experiments demonstrated that one algorithm improved over other approaches.
彩色滤波阵列(CFA)在数码相机中得到了广泛的应用。在文献中有许多变体的cfa。最近,我们提出了一种新的CFA,称为CFA 3.0,与一些标准的CFA相比,CFA 3.0具有合理的性能。在本文中,我们研究了使用图像修复算法来进一步提高CFA 3.0的去马赛克性能。比较了六种传统的和基于深度学习的图像绘制算法。大量的实验表明,其中一种算法优于其他方法。
{"title":"Further Improvements of CFA 3.0 by Combining Inpainting and Pansharpening Techniques","authors":"C. Kwan, Jude Larkin","doi":"10.5121/SIPIJ.2020.11601","DOIUrl":"https://doi.org/10.5121/SIPIJ.2020.11601","url":null,"abstract":"Color Filter Array (CFA) has been widely used in digital cameras. There are many variants of CFAs in the literature. Recently, a new CFA known as CFA 3.0 was proposed by us and has been shown to yield reasonable performance as compared to some standard ones. In this paper, we investigate the use of inpainting algorithms to further improve the demosaicing performance of CFA 3.0. Six conventional and deep learning based inpainting algorithms were compared. Extensive experiments demonstrated that one algorithm improved over other approaches.","PeriodicalId":90726,"journal":{"name":"Signal and image processing : an international journal","volume":"7 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75454891","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Eye Gaze Estimation Invisible and IR Spectrum for Driver Monitoring System 驾驶员监控系统的眼注视估计、不可见光谱和红外光谱
Pub Date : 2020-10-30 DOI: 10.5121/sipij.2020.11501
Susmitha Mohan, M. Phirke
Driver monitoring system has gained lot of popularity in automotive sector to ensure safety while driving. Collisions due to driver inattentiveness or driver fatigue or over reliance on autonomous driving features arethe major reasons for road accidents and fatalities. Driver monitoring systems aims to monitor various aspect of driving and provides appropriate warnings whenever required. Eye gaze estimation is a key element in almost all of the driver monitoring systems. Gaze estimation aims to find the point of gaze which is basically,” -where is driver looking”. This helps in understanding if the driver is attentively looking at the road or if he is distracted. Estimating gaze point also plays important role in many other applications like retail shopping, online marketing, psychological tests, healthcare etc. This paper covers the various aspects of eye gaze estimation for a driver monitoring system including sensor choice and sensor placement. There are multiple ways by which eye gaze estimation can be done. A detailed comparative study on two of the popular methods for gaze estimation using eye features is covered in this paper. An infra-red camera is used to capture data for this study. Method 1 tracks corneal reflection centre w.r.t the pupil centre and method 2 tracks the pupil centre w.r.t the eye centre to estimate gaze. There are advantages and disadvantages with both the methods which has been looked into. This paper can act as a reference for researchers working in the same field to understand possibilities and limitations of eye gaze estimation for driver monitoring system.
驾驶员监控系统在汽车行业得到了广泛的应用,以确保驾驶安全。由于驾驶员注意力不集中或驾驶员疲劳或过度依赖自动驾驶功能而导致的碰撞是道路交通事故和死亡的主要原因。驾驶员监控系统旨在监控驾驶的各个方面,并在需要时提供适当的警告。人眼注视估计是几乎所有驾驶员监控系统中的一个关键因素。凝视估计的目的是找到凝视点,基本上就是“司机在看哪里”。这有助于了解司机是否在专心看路,还是在分心。估计凝视点在零售购物、在线营销、心理测试、医疗保健等许多其他应用中也发挥着重要作用。本文涵盖了驾驶员监控系统中人眼注视估计的各个方面,包括传感器的选择和传感器的放置。有多种方法可以完成眼睛注视的估计。本文对两种常用的基于人眼特征的注视估计方法进行了详细的比较研究。红外摄像机被用来捕捉本研究的数据。方法1以瞳孔中心为中心跟踪角膜反射中心,方法2以眼中心为中心跟踪瞳孔中心估计凝视。这两种方法都有各自的优点和缺点。本文可以为同行研究人员了解人眼注视估计在驾驶员监控系统中的可能性和局限性提供参考。
{"title":"Eye Gaze Estimation Invisible and IR Spectrum for Driver Monitoring System","authors":"Susmitha Mohan, M. Phirke","doi":"10.5121/sipij.2020.11501","DOIUrl":"https://doi.org/10.5121/sipij.2020.11501","url":null,"abstract":"Driver monitoring system has gained lot of popularity in automotive sector to ensure safety while driving. Collisions due to driver inattentiveness or driver fatigue or over reliance on autonomous driving features arethe major reasons for road accidents and fatalities. Driver monitoring systems aims to monitor various aspect of driving and provides appropriate warnings whenever required. Eye gaze estimation is a key element in almost all of the driver monitoring systems. Gaze estimation aims to find the point of gaze which is basically,” -where is driver looking”. This helps in understanding if the driver is attentively looking at the road or if he is distracted. Estimating gaze point also plays important role in many other applications like retail shopping, online marketing, psychological tests, healthcare etc. This paper covers the various aspects of eye gaze estimation for a driver monitoring system including sensor choice and sensor placement. There are multiple ways by which eye gaze estimation can be done. A detailed comparative study on two of the popular methods for gaze estimation using eye features is covered in this paper. An infra-red camera is used to capture data for this study. Method 1 tracks corneal reflection centre w.r.t the pupil centre and method 2 tracks the pupil centre w.r.t the eye centre to estimate gaze. There are advantages and disadvantages with both the methods which has been looked into. This paper can act as a reference for researchers working in the same field to understand possibilities and limitations of eye gaze estimation for driver monitoring system.","PeriodicalId":90726,"journal":{"name":"Signal and image processing : an international journal","volume":"73 1","pages":"1-20"},"PeriodicalIF":0.0,"publicationDate":"2020-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75754908","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Face Verification Across Age Progression using Enhanced Convolution Neural Network 基于增强卷积神经网络的跨年龄人脸验证
Pub Date : 2020-10-30 DOI: 10.5121/sipij.2020.11504
A. M. Osman, Serestina Viriri
This paper proposes a deep learning method for facial verification of aging subjects. Facial aging is a texture and shape variations that affect the human face as time progresses. Accordingly, there is a demand to develop robust methods to verify facial images when they age. In this paper, a deep learning method based on GoogLeNet pre-trained convolution network fused with Histogram Orientation Gradient (HOG) and Local Binary Pattern (LBP) feature descriptors have been applied for feature extraction and classification. The experiments are based on the facial images collected from MORPH and FG-Net benchmarked datasets. Euclidean distance has been used to measure the similarity between pairs of feature vectors with the age gap. Experiments results show an improvement in the validation accuracy conducted on the FG-NET database, which it reached 100%, while with MORPH database the validation accuracy is 99.8%. The proposed method has better performance and higher accuracy than current state-of-the-art methods.
提出了一种基于深度学习的人脸识别方法。面部老化是一种随着时间的推移而影响人脸的纹理和形状变化。因此,有必要开发强大的方法来验证面部图像时,他们的年龄。本文将基于GoogLeNet预训练卷积网络的深度学习方法与直方图方向梯度(Histogram Orientation Gradient, HOG)和局部二值模式(Local Binary Pattern, LBP)特征描述子相融合,用于特征提取和分类。实验基于MORPH和FG-Net基准数据集收集的面部图像。欧几里得距离被用来衡量年龄差距对特征向量之间的相似性。实验结果表明,FG-NET数据库的验证准确率达到100%,而MORPH数据库的验证准确率为99.8%。与现有方法相比,该方法具有更好的性能和更高的精度。
{"title":"Face Verification Across Age Progression using Enhanced Convolution Neural Network","authors":"A. M. Osman, Serestina Viriri","doi":"10.5121/sipij.2020.11504","DOIUrl":"https://doi.org/10.5121/sipij.2020.11504","url":null,"abstract":"This paper proposes a deep learning method for facial verification of aging subjects. Facial aging is a texture and shape variations that affect the human face as time progresses. Accordingly, there is a demand to develop robust methods to verify facial images when they age. In this paper, a deep learning method based on GoogLeNet pre-trained convolution network fused with Histogram Orientation Gradient (HOG) and Local Binary Pattern (LBP) feature descriptors have been applied for feature extraction and classification. The experiments are based on the facial images collected from MORPH and FG-Net benchmarked datasets. Euclidean distance has been used to measure the similarity between pairs of feature vectors with the age gap. Experiments results show an improvement in the validation accuracy conducted on the FG-NET database, which it reached 100%, while with MORPH database the validation accuracy is 99.8%. The proposed method has better performance and higher accuracy than current state-of-the-art methods.","PeriodicalId":90726,"journal":{"name":"Signal and image processing : an international journal","volume":"40 5 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88510441","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Batch Normalized Convolution Neural Network for Liver Segmentation 批量归一化卷积神经网络肝脏分割
Pub Date : 2020-10-30 DOI: 10.5121/sipij.2020.11502
Fatima Abdalbagi, Serestina Viriri, M. T. Mohammed
With the huge innovative improvement in all lifestyles, it has been important to build up the clinical fields, remembering the finding for which treatment is done; where the fruitful treatment relies upon the preoperative. Models for the preoperative, for example, planning to understand the complex internal structure of the liver and precisely localize the liver surface and its tumors; there are various algorithms proposed to do the automatic liver segmentation. In this paper, we propose a Batch Normalization After All Convolutional Neural Network (BATA-Convnet) model to segment the liver CT images using Deep Learning Technique. The proposed liver segmentation model consists of four main steps: pre-processing, training the BATA-Convnet, liver segmentation, and the postprocessing step to maximize the result efficiency. Medical Image Computing and Computer Assisted Intervention (MICCAI) dataset and 3DImage Reconstruction for Comparison of Algorithm Database (3D-IRCAD) were used in the experimentation and the average results using MICCAI are 0.91% for Dice, 13.44% for VOE, 0.23% for RVD, 0.29mm for ASD, 1.35mm for RMSSD and 0.36mm for MaxASD. The average results using 3DIRCAD dataset are 0.84% for Dice, 13.24% for VOE, 0.16% for RVD, 0.32mm for ASD, 1.17mm for RMSSD and 0.33mm for MaxASD.
随着所有生活方式的巨大创新改善,建立临床领域变得非常重要,记住治疗的发现;有效的治疗依赖于术前。用于术前的模型,例如,计划了解肝脏复杂的内部结构并精确定位肝脏表面及其肿瘤;目前已经提出了多种算法来实现自动肝脏分割。在本文中,我们提出了一种Batch Normalization After All Convolutional Neural Network (BATA-Convnet)模型,利用深度学习技术对肝脏CT图像进行分割。本文提出的肝分割模型主要包括预处理、训练BATA-Convnet、肝分割和最大化结果效率的后处理四个步骤。使用医学图像计算与计算机辅助干预(MICCAI)数据集和3d图像重建算法对比数据库(3D-IRCAD)进行实验,MICCAI的平均结果为Dice的0.91%,VOE的13.44%,RVD的0.23%,ASD的0.29mm, RMSSD的1.35mm和MaxASD的0.36mm。使用3DIRCAD数据集的平均结果是Dice为0.84%,VOE为13.24%,RVD为0.16%,ASD为0.32mm, RMSSD为1.17mm, MaxASD为0.33mm。
{"title":"Batch Normalized Convolution Neural Network for Liver Segmentation","authors":"Fatima Abdalbagi, Serestina Viriri, M. T. Mohammed","doi":"10.5121/sipij.2020.11502","DOIUrl":"https://doi.org/10.5121/sipij.2020.11502","url":null,"abstract":"With the huge innovative improvement in all lifestyles, it has been important to build up the clinical fields, remembering the finding for which treatment is done; where the fruitful treatment relies upon the preoperative. Models for the preoperative, for example, planning to understand the complex internal structure of the liver and precisely localize the liver surface and its tumors; there are various algorithms proposed to do the automatic liver segmentation. In this paper, we propose a Batch Normalization After All Convolutional Neural Network (BATA-Convnet) model to segment the liver CT images using Deep Learning Technique. The proposed liver segmentation model consists of four main steps: pre-processing, training the BATA-Convnet, liver segmentation, and the postprocessing step to maximize the result efficiency. Medical Image Computing and Computer Assisted Intervention (MICCAI) dataset and 3DImage Reconstruction for Comparison of Algorithm Database (3D-IRCAD) were used in the experimentation and the average results using MICCAI are 0.91% for Dice, 13.44% for VOE, 0.23% for RVD, 0.29mm for ASD, 1.35mm for RMSSD and 0.36mm for MaxASD. The average results using 3DIRCAD dataset are 0.84% for Dice, 13.24% for VOE, 0.16% for RVD, 0.32mm for ASD, 1.17mm for RMSSD and 0.33mm for MaxASD.","PeriodicalId":90726,"journal":{"name":"Signal and image processing : an international journal","volume":"4 1","pages":"21-35"},"PeriodicalIF":0.0,"publicationDate":"2020-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72873807","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Gender Discrimination based on the Thermal Signature of the Face and the External Ear 基于面部和外耳热特征的性别歧视
Pub Date : 2020-08-31 DOI: 10.5121/sipij.2020.11402
G. Koukiou, V. Anastassopoulos
Simple features extracted from the thermal infrared images of the persons' face are proposed for gender discrimination. Two different types of thermal features are used. The first type is actually based on the mean value of the pixels of specific locations on the face. All cases of persons from the used database, males and females, are correctly distinguished based on this feature. Classification results are verified using two conventional approaches, namely: a. the simplest possible neural network so that generalization is achieved along with successful discrimination between all persons and b. the leave-one-out approach to demonstrate the classification performance on unknown persons using the simplest classifiers possible. The second type takes advantage of the temperature distribution on the ear of the persons. It is found that for the men the cooler region on the ear is larger as percentage compared to that of the women.
提出了从人脸热红外图像中提取简单特征进行性别歧视的方法。使用了两种不同类型的热特性。第一种类型实际上是基于面部特定位置像素的平均值。基于这一特征,可以正确区分数据库中所有的男性和女性病例。使用两种常规方法验证分类结果,即:a.尽可能简单的神经网络,从而实现泛化并成功区分所有人;b.使用尽可能简单的分类器来展示对未知人的分类性能的留一方法。第二种是利用人耳上的温度分布。研究发现,男性耳朵上较冷的区域所占的比例要比女性大。
{"title":"Gender Discrimination based on the Thermal Signature of the Face and the External Ear","authors":"G. Koukiou, V. Anastassopoulos","doi":"10.5121/sipij.2020.11402","DOIUrl":"https://doi.org/10.5121/sipij.2020.11402","url":null,"abstract":"Simple features extracted from the thermal infrared images of the persons' face are proposed for gender discrimination. Two different types of thermal features are used. The first type is actually based on the mean value of the pixels of specific locations on the face. All cases of persons from the used database, males and females, are correctly distinguished based on this feature. Classification results are verified using two conventional approaches, namely: a. the simplest possible neural network so that generalization is achieved along with successful discrimination between all persons and b. the leave-one-out approach to demonstrate the classification performance on unknown persons using the simplest classifiers possible. The second type takes advantage of the temperature distribution on the ear of the persons. It is found that for the men the cooler region on the ear is larger as percentage compared to that of the women.","PeriodicalId":90726,"journal":{"name":"Signal and image processing : an international journal","volume":"65 1","pages":"13-23"},"PeriodicalIF":0.0,"publicationDate":"2020-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83171639","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
Signal and image processing : an international journal
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1