首页 > 最新文献

Proceedings of the 4th International Conference on Image Processing and Machine Vision最新文献

英文 中文
CLAMOT: 3D Detection and Tracking via Multi-modal Feature Aggregation CLAMOT:基于多模态特征聚合的三维检测和跟踪
Shuo Zhang, Xiaolong Liu, Wenqi Tao
In autonomous driving, multi-object tracking (MOT) can help vehicles perceive surroundings better and perform well-informed motion-planning. Methods based on LiDAR suffer from the sparsity of LiDAR points and detect only in a limited range. To this end, we propose a camera and LiDAR aggregation module named CLA-fusion to fuse the two modal features in a point-wise manner. The enhanced points can be used for extracting features through a 3D backbone. For the detection, we adopts a center-based method which means detecting the centers of objects by a keypoint detector and regressing other attributes, like 3D size, velocity, etc. In the tracking part, we use a simple but effective matching strategy, closest-point matching. According to the structure and characteristics of the whole framework, we name our model CLAMOT. Our experiments on nuScenes and Waymo benchmarks achieve competitive results.
在自动驾驶中,多目标跟踪(MOT)可以帮助车辆更好地感知周围环境并执行明智的运动规划。基于激光雷达的方法受到激光雷达点的稀疏性和探测范围的限制。为此,我们提出了一个名为CLA-fusion的相机和LiDAR聚合模块,以点为方向融合两个模态特征。增强点可用于通过三维主干提取特征。对于检测,我们采用基于中心的方法,即通过关键点检测器检测物体的中心,并回归其他属性,如3D尺寸,速度等。在跟踪部分,我们使用了一种简单而有效的匹配策略——最近点匹配。根据整个框架的结构和特点,我们将模型命名为CLAMOT。我们在nuScenes和Waymo基准上的实验取得了具有竞争力的结果。
{"title":"CLAMOT: 3D Detection and Tracking via Multi-modal Feature Aggregation","authors":"Shuo Zhang, Xiaolong Liu, Wenqi Tao","doi":"10.1145/3529446.3529451","DOIUrl":"https://doi.org/10.1145/3529446.3529451","url":null,"abstract":"In autonomous driving, multi-object tracking (MOT) can help vehicles perceive surroundings better and perform well-informed motion-planning. Methods based on LiDAR suffer from the sparsity of LiDAR points and detect only in a limited range. To this end, we propose a camera and LiDAR aggregation module named CLA-fusion to fuse the two modal features in a point-wise manner. The enhanced points can be used for extracting features through a 3D backbone. For the detection, we adopts a center-based method which means detecting the centers of objects by a keypoint detector and regressing other attributes, like 3D size, velocity, etc. In the tracking part, we use a simple but effective matching strategy, closest-point matching. According to the structure and characteristics of the whole framework, we name our model CLAMOT. Our experiments on nuScenes and Waymo benchmarks achieve competitive results.","PeriodicalId":151062,"journal":{"name":"Proceedings of the 4th International Conference on Image Processing and Machine Vision","volume":"109 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133819605","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DnT: Learning Unsupervised Denoising Transformer from Single Noisy Image DnT:从单个噪声图像中学习无监督去噪变压器
Xiaolong Liu, Yu Hong, Qifang Yin, Shuo Zhang
In the last few years, a myriad of Transformer based methods have drawn considerable attention due to their outstanding performance on various computer vision tasks. However, most image denoising methods are based on convolutional neural networks (CNNs), few attempts have been made with Transformer, especially in self-supervised and unsupervised methods. In this paper, we propose a novel and good performance unsupervised image Denoising Transformer (DnT) which is just trained by the single input noisy image. Our network combines Transformer and CNN to predict the counterpart clean target, the training loss was measured by pairs of noisy independent images constructed from the input image. The dropout-based ensemble is used to get the final denoised result by averaging multiple predictions generated by the trained model. Experiments show that the proposed method not only has superior performance over the state-of-the-art single noisy image denoiser on additive white Gaussian noise (AWGN) removal but also achieves good results on real-world image denoising.
在过去的几年中,无数基于Transformer的方法由于其在各种计算机视觉任务上的出色表现而引起了相当大的关注。然而,大多数图像去噪方法都是基于卷积神经网络(cnn),很少有人尝试使用Transformer,特别是在自监督和无监督方法中。本文提出了一种新型的、性能良好的无监督图像去噪变压器(DnT),该变压器仅由单输入噪声图像进行训练。我们的网络结合了Transformer和CNN来预测对应的干净目标,训练损失是由输入图像构建的独立噪声图像对来测量的。基于dropout的集成通过对训练模型产生的多个预测进行平均,得到最终去噪结果。实验表明,该方法不仅在去除加性高斯白噪声(AWGN)方面优于现有的单噪声图像去噪方法,而且在实际图像去噪方面也取得了良好的效果。
{"title":"DnT: Learning Unsupervised Denoising Transformer from Single Noisy Image","authors":"Xiaolong Liu, Yu Hong, Qifang Yin, Shuo Zhang","doi":"10.1145/3529446.3529455","DOIUrl":"https://doi.org/10.1145/3529446.3529455","url":null,"abstract":"In the last few years, a myriad of Transformer based methods have drawn considerable attention due to their outstanding performance on various computer vision tasks. However, most image denoising methods are based on convolutional neural networks (CNNs), few attempts have been made with Transformer, especially in self-supervised and unsupervised methods. In this paper, we propose a novel and good performance unsupervised image Denoising Transformer (DnT) which is just trained by the single input noisy image. Our network combines Transformer and CNN to predict the counterpart clean target, the training loss was measured by pairs of noisy independent images constructed from the input image. The dropout-based ensemble is used to get the final denoised result by averaging multiple predictions generated by the trained model. Experiments show that the proposed method not only has superior performance over the state-of-the-art single noisy image denoiser on additive white Gaussian noise (AWGN) removal but also achieves good results on real-world image denoising.","PeriodicalId":151062,"journal":{"name":"Proceedings of the 4th International Conference on Image Processing and Machine Vision","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120954911","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Adaptive Covariance Matrix based on Blur Evaluation for Visual-Inertial Navigation 基于模糊评价的自适应协方差矩阵视觉惯性导航
Yihao Zuo, C. Yan, Qiwei Liu, Xia Wang
The covariance matrix in the current mainstream visual-inertial navigation system is artificially set and the weight of visual information cannot be adjusted by different blur degree, which cause the poor accuracy and robustness in the whole system. In order to solve this problem, this paper proposed a navigation scheme based on adaptive covariance matrix. This method used the Laplacian operator to evaluate the blur degree of image by a score. And then the visual covariance matrix is adjusted according to the different scores, which can adjust the weight in the fusion system according to the image quality. By doing this, the algorithm can improve the accuracy of the system. The simulation results show that the proposed method can effectively improve the system accuracy. Compared with the traditional method, the proposed algorithm has stronger robustness when motion blur occur.
当前主流视觉惯性导航系统的协方差矩阵是人为设置的,不能通过不同的模糊程度来调整视觉信息的权重,导致整个系统的精度和鲁棒性较差。为了解决这一问题,本文提出了一种基于自适应协方差矩阵的导航方案。该方法采用拉普拉斯算子对图像的模糊程度进行评分。然后根据不同的分数调整视觉协方差矩阵,可以根据图像质量调整融合系统中的权重。通过这样做,该算法可以提高系统的精度。仿真结果表明,该方法能有效提高系统精度。与传统方法相比,该算法在运动模糊情况下具有更强的鲁棒性。
{"title":"Adaptive Covariance Matrix based on Blur Evaluation for Visual-Inertial Navigation","authors":"Yihao Zuo, C. Yan, Qiwei Liu, Xia Wang","doi":"10.1145/3529446.3529462","DOIUrl":"https://doi.org/10.1145/3529446.3529462","url":null,"abstract":"The covariance matrix in the current mainstream visual-inertial navigation system is artificially set and the weight of visual information cannot be adjusted by different blur degree, which cause the poor accuracy and robustness in the whole system. In order to solve this problem, this paper proposed a navigation scheme based on adaptive covariance matrix. This method used the Laplacian operator to evaluate the blur degree of image by a score. And then the visual covariance matrix is adjusted according to the different scores, which can adjust the weight in the fusion system according to the image quality. By doing this, the algorithm can improve the accuracy of the system. The simulation results show that the proposed method can effectively improve the system accuracy. Compared with the traditional method, the proposed algorithm has stronger robustness when motion blur occur.","PeriodicalId":151062,"journal":{"name":"Proceedings of the 4th International Conference on Image Processing and Machine Vision","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114386906","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Seismic Data Interpolation by the Projected Iterative Soft-threshold Algorithm for Tight Frame 基于投影迭代软阈值算法的紧框架地震数据插值
Lin Tian, S. Qin
Seismic data recovery from missing traces is a crucial step in seismic data pre-processing. Recently researches have proposed many useful methods to reconstruct the seismic data based on compressed sensing. Curvelet frames can be used to sparsely represent the seismic data volume, analysis model has been proposed to reconstruct the seismic data, however, the latest kind of discrete curvelet transform has tight frame property, the recent insights show synthetically model is more suitable for a tight frame. A synthetically model is introduced to seismic data reconstruction; projected iterative soft-threshold algorithm (pFISTA) is used to solve the model. The recovery performs well on synthetic as well as real data by the proposed method. Comparing with the analysis model solved by an iterative soft-threshold algorithm (FISTA) in the curvelet domain, the new method has improved reconstruction efficiency and reduced the computation time.
地震资料失道恢复是地震资料预处理的关键步骤。近年来的研究提出了许多有用的基于压缩感知的地震数据重建方法。曲线框架可以稀疏地表示地震数据体,人们提出了分析模型来重建地震数据,但最新的离散曲线转换具有紧框架性质,最近的见解表明综合模型更适合于紧框架。将综合模型引入到地震数据重建中;采用投影迭代软阈值算法(pFISTA)求解模型。该方法对合成数据和实际数据均有较好的恢复效果。与曲线域迭代软阈值算法(FISTA)求解的分析模型相比,该方法提高了重建效率,减少了计算时间。
{"title":"Seismic Data Interpolation by the Projected Iterative Soft-threshold Algorithm for Tight Frame","authors":"Lin Tian, S. Qin","doi":"10.1145/3529446.3529460","DOIUrl":"https://doi.org/10.1145/3529446.3529460","url":null,"abstract":"Seismic data recovery from missing traces is a crucial step in seismic data pre-processing. Recently researches have proposed many useful methods to reconstruct the seismic data based on compressed sensing. Curvelet frames can be used to sparsely represent the seismic data volume, analysis model has been proposed to reconstruct the seismic data, however, the latest kind of discrete curvelet transform has tight frame property, the recent insights show synthetically model is more suitable for a tight frame. A synthetically model is introduced to seismic data reconstruction; projected iterative soft-threshold algorithm (pFISTA) is used to solve the model. The recovery performs well on synthetic as well as real data by the proposed method. Comparing with the analysis model solved by an iterative soft-threshold algorithm (FISTA) in the curvelet domain, the new method has improved reconstruction efficiency and reduced the computation time.","PeriodicalId":151062,"journal":{"name":"Proceedings of the 4th International Conference on Image Processing and Machine Vision","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133596458","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Infrared small target detection algorithm with complex background based on YOLO-NWD 基于YOLO-NWD的复杂背景红外小目标检测算法
Xiao Zhou, Lang Jiang, Xujun Guan, Xingang Mou
Because of small number of occupied pixels, lacking shape and texture information, the reliability of infrared remote target detection has always been a difficult research topic. To improve the accuracy and precision of detection of infrared small targets under complex background conditions, a deep learning-based infrared small target detection algorithm YOLO-NWD is proposed. According to the characteristics of small and medium targets in infrared images, multi-channel feature fusion image was used as the input of YOLO detection framework combined with image preprocessing method. Combined with SE module and ASPP module, feature weights are explored to improve feature utilization efficiency. Finally, the normalized Wasserstein distance (NWD) loss is used to replace the original IoU calculation loss to reduce the sensitivity of small target position deviation. The experimental results show that the algorithm proposed in this paper improves the accuracy by 2.5% and the recall rate by 4%.
红外遥感目标检测的可靠性一直是红外遥感目标检测的一个难点,因为红外遥感目标被占用的像元数量少,缺乏形状和纹理信息。为了提高复杂背景条件下红外小目标检测的准确度和精度,提出了一种基于深度学习的红外小目标检测算法YOLO-NWD。根据红外图像中中小目标的特点,结合图像预处理方法,采用多通道特征融合图像作为YOLO检测框架的输入。结合SE模块和ASPP模块,探索特征权重,提高特征利用效率。最后,利用归一化Wasserstein距离(NWD)损失代替原有IoU计算损失,降低目标位置小偏差的敏感性。实验结果表明,本文算法的准确率提高了2.5%,召回率提高了4%。
{"title":"Infrared small target detection algorithm with complex background based on YOLO-NWD","authors":"Xiao Zhou, Lang Jiang, Xujun Guan, Xingang Mou","doi":"10.1145/3529446.3529448","DOIUrl":"https://doi.org/10.1145/3529446.3529448","url":null,"abstract":"Because of small number of occupied pixels, lacking shape and texture information, the reliability of infrared remote target detection has always been a difficult research topic. To improve the accuracy and precision of detection of infrared small targets under complex background conditions, a deep learning-based infrared small target detection algorithm YOLO-NWD is proposed. According to the characteristics of small and medium targets in infrared images, multi-channel feature fusion image was used as the input of YOLO detection framework combined with image preprocessing method. Combined with SE module and ASPP module, feature weights are explored to improve feature utilization efficiency. Finally, the normalized Wasserstein distance (NWD) loss is used to replace the original IoU calculation loss to reduce the sensitivity of small target position deviation. The experimental results show that the algorithm proposed in this paper improves the accuracy by 2.5% and the recall rate by 4%.","PeriodicalId":151062,"journal":{"name":"Proceedings of the 4th International Conference on Image Processing and Machine Vision","volume":"82 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117250498","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
YOLO-oil: A Real-time Transformer Fault Detector toward Small Dataset 面向小数据集的实时变压器故障检测器YOLO-oil
Shaojie Hu, Xianwen Jin, Huigang Wang
In electrical substations, fault detection of the transformer widely relies on human eye, which is low efficiency and costly. With under-oil robot and deep learning algorithms, the fault detection will be done without draining the transformer oil. As we know, deep learning methods for computer vision have achieved incredible results on some tasks such as object detection. However, such success greatly relies on the huge dataset, which is extremely high-cost and unavailable in some industry application. Deep learning algorithm, such as YOLO series, often fails on small dataset, and the test accuracy decreases significantly due to the neural network overfitting on the small dataset. In this paper, the YOLO-oil network for transformer fault detector based on YOLOv5 is proposed to mitigate the overfitting problem on small dataset: First, we shrink the network depth and get a light weight backbone. Second, we improved the network architecture by decoupling the detect head network. Since no open dataset exists for transformer fault detection before, the author creates a brand-new training dataset and a test dataset. Experimental results on the test set show that our algorithm achieves surprising results for the transformer fault detection task and surpasses YOLOv5, which is a great help to industry application.
在变电站中,变压器故障检测普遍依赖人眼,效率低,成本高。利用油下机器人和深度学习算法,可以在不消耗变压器油的情况下完成故障检测。正如我们所知,计算机视觉的深度学习方法在一些任务上取得了令人难以置信的结果,比如物体检测。然而,这样的成功很大程度上依赖于庞大的数据集,而这些数据集的成本极高,在一些行业应用中是不可用的。深度学习算法,如YOLO系列,在小数据集上经常失败,并且由于神经网络在小数据集上的过拟合导致测试精度显著下降。针对小数据集上的过拟合问题,提出了基于YOLOv5的变压器故障检测YOLO-oil网络:首先,我们缩小网络深度,得到一个轻量级的骨干网络;其次,我们通过解耦检测头网络来改进网络结构。由于以前没有开放的变压器故障检测数据集,作者创建了一个全新的训练数据集和测试数据集。在测试集上的实验结果表明,我们的算法在变压器故障检测任务上取得了惊人的效果,并且超过了YOLOv5,对工业应用有很大的帮助。
{"title":"YOLO-oil: A Real-time Transformer Fault Detector toward Small Dataset","authors":"Shaojie Hu, Xianwen Jin, Huigang Wang","doi":"10.1145/3529446.3529461","DOIUrl":"https://doi.org/10.1145/3529446.3529461","url":null,"abstract":"In electrical substations, fault detection of the transformer widely relies on human eye, which is low efficiency and costly. With under-oil robot and deep learning algorithms, the fault detection will be done without draining the transformer oil. As we know, deep learning methods for computer vision have achieved incredible results on some tasks such as object detection. However, such success greatly relies on the huge dataset, which is extremely high-cost and unavailable in some industry application. Deep learning algorithm, such as YOLO series, often fails on small dataset, and the test accuracy decreases significantly due to the neural network overfitting on the small dataset. In this paper, the YOLO-oil network for transformer fault detector based on YOLOv5 is proposed to mitigate the overfitting problem on small dataset: First, we shrink the network depth and get a light weight backbone. Second, we improved the network architecture by decoupling the detect head network. Since no open dataset exists for transformer fault detection before, the author creates a brand-new training dataset and a test dataset. Experimental results on the test set show that our algorithm achieves surprising results for the transformer fault detection task and surpasses YOLOv5, which is a great help to industry application.","PeriodicalId":151062,"journal":{"name":"Proceedings of the 4th International Conference on Image Processing and Machine Vision","volume":"76 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127506212","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Infrared dim and small target detection based on total variation and multiple noise constraints modeling 基于全变分和多噪声约束建模的红外弱小目标检测
Xiaowen Wang, Xiaoyan Xia, Qiao Li, Wei Xue
To improve the ability of infrared dim small target detection algorithm based on traditional infrared patch-image (IPI) model, a new detection model based on total variation and multiple noise constraints is proposed. We firstly transform the original infrared image into an IPI, and then the total variational regularization constrains the background patch-image in order to reduce the noise on the target image. In the meantime, the edge information of the image can be preserved to avoid excessive smoothness of the restored background image. Additionally, considering the lack of noise distribution in the patch-image, the combined and norm are introduced to describe the noise more accurately. The experimental results show that the proposed method can suppress the background clutter better and improve detection performance effectively.
为了提高基于传统红外补丁图像(IPI)模型的红外弱小目标检测算法的能力,提出了一种基于全变分和多噪声约束的红外弱小目标检测模型。首先将原始红外图像转换为IPI,然后对背景块图像进行全变分正则化约束,以降低目标图像上的噪声。同时,可以保留图像的边缘信息,避免恢复后的背景图像过于平滑。此外,考虑到图像中噪声分布的不足,引入了组合范数来更准确地描述噪声。实验结果表明,该方法能较好地抑制背景杂波,有效提高检测性能。
{"title":"Infrared dim and small target detection based on total variation and multiple noise constraints modeling","authors":"Xiaowen Wang, Xiaoyan Xia, Qiao Li, Wei Xue","doi":"10.1145/3529446.3529447","DOIUrl":"https://doi.org/10.1145/3529446.3529447","url":null,"abstract":"To improve the ability of infrared dim small target detection algorithm based on traditional infrared patch-image (IPI) model, a new detection model based on total variation and multiple noise constraints is proposed. We firstly transform the original infrared image into an IPI, and then the total variational regularization constrains the background patch-image in order to reduce the noise on the target image. In the meantime, the edge information of the image can be preserved to avoid excessive smoothness of the restored background image. Additionally, considering the lack of noise distribution in the patch-image, the combined and norm are introduced to describe the noise more accurately. The experimental results show that the proposed method can suppress the background clutter better and improve detection performance effectively.","PeriodicalId":151062,"journal":{"name":"Proceedings of the 4th International Conference on Image Processing and Machine Vision","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126858767","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Group Sparse-based Discriminative Feature Learning for Face Recognition 基于组稀疏的判别特征学习人脸识别
Xiaoqun Qiu, Xiaoyu Du, Liyan Deng, Zhen Chen
The rapid development of facial recognition technology has brought great convenience to daily life, but also serious security risks, especially in the case of occlusion and loud noise. Faced with this limitation, this letter proposes a fast face recognition framework called a Group Sparse-based Discriminative Feature Learning (GSDFL-Net). Specifically, GSDFL-Net uses a novel unified objective function to simultaneously learn the discriminant features, sparse code and classification errors. In the proposed framework, the feature projection is incorporated into GSDFL-Net model, which reduces the classification errors. Then, we integrate denoising FFDNet into the proposed GS FL-Net model to penalize the noisy pixels, which is simultaneously learned by our unified objective function. Besides, we derive an optimization mechanism to encourage obtained learning parameters and decrease the information loss. Extensive experiments demonstrate the effectiveness of the proposed scheme under different including occlusion random noise conditions on the famous Aleix Martinez and ExYale B database.
人脸识别技术的快速发展给日常生活带来了极大的便利,但也带来了严重的安全隐患,特别是在遮挡和噪声较大的情况下。面对这一限制,本文提出了一种快速人脸识别框架,称为基于组稀疏的判别特征学习(GSDFL-Net)。具体来说,GSDFL-Net使用一种新颖的统一目标函数来同时学习判别特征、稀疏代码和分类错误。在该框架中,将特征投影纳入GSDFL-Net模型,降低了分类误差。然后,我们将去噪FFDNet整合到所提出的GS FL-Net模型中,对噪声像素进行惩罚,同时通过我们统一的目标函数进行学习。此外,我们还推导了一种优化机制,以鼓励获得的学习参数,减少信息损失。在著名的Aleix Martinez和ExYale B数据库上进行的大量实验证明了该方案在不同包括遮挡随机噪声条件下的有效性。
{"title":"Group Sparse-based Discriminative Feature Learning for Face Recognition","authors":"Xiaoqun Qiu, Xiaoyu Du, Liyan Deng, Zhen Chen","doi":"10.1145/3529446.3529450","DOIUrl":"https://doi.org/10.1145/3529446.3529450","url":null,"abstract":"The rapid development of facial recognition technology has brought great convenience to daily life, but also serious security risks, especially in the case of occlusion and loud noise. Faced with this limitation, this letter proposes a fast face recognition framework called a Group Sparse-based Discriminative Feature Learning (GSDFL-Net). Specifically, GSDFL-Net uses a novel unified objective function to simultaneously learn the discriminant features, sparse code and classification errors. In the proposed framework, the feature projection is incorporated into GSDFL-Net model, which reduces the classification errors. Then, we integrate denoising FFDNet into the proposed GS FL-Net model to penalize the noisy pixels, which is simultaneously learned by our unified objective function. Besides, we derive an optimization mechanism to encourage obtained learning parameters and decrease the information loss. Extensive experiments demonstrate the effectiveness of the proposed scheme under different including occlusion random noise conditions on the famous Aleix Martinez and ExYale B database.","PeriodicalId":151062,"journal":{"name":"Proceedings of the 4th International Conference on Image Processing and Machine Vision","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126629022","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Proceedings of the 4th International Conference on Image Processing and Machine Vision 第四届图像处理与机器视觉国际会议论文集
{"title":"Proceedings of the 4th International Conference on Image Processing and Machine Vision","authors":"","doi":"10.1145/3529446","DOIUrl":"https://doi.org/10.1145/3529446","url":null,"abstract":"","PeriodicalId":151062,"journal":{"name":"Proceedings of the 4th International Conference on Image Processing and Machine Vision","volume":"40 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131137859","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the 4th International Conference on Image Processing and Machine Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1