首页 > 最新文献

Proceedings of the 2022 11th International Conference on Computing and Pattern Recognition最新文献

英文 中文
High-speed Video Measurement Method of Attitude Angle Applied to Floating Offshore Wind Turbine Model Wind-Wave Tank Test 浮式海上风电模型风浪箱试验中姿态角高速视频测量方法
Qing Zhong, Jiahao Liu, Peng Chen
During the wind-wave tank model test of floating offshore wind turbine (FOWT), acquiring the model attitude angles is one of the key objectives. Due to the environment of the wind-wave tank, there are inevitable shortcomings of traditional contact sensors, such as difficult installation, increased model weight, limited sampling frequency, and incomplete data acquisition. In this paper, a vision-based technology is presented, which mainly consists of (1) construct a high-speed video acquisition network; (2) identify, track and match the circular landmarks on the sequence image, and obtain three-dimensional coordinates of the spatial points using the spatial point reconstruction algorithm; (3) propose a method for computing model attitude angles in sequence images to obtain the pose of the model in each frame; (4) deduce the theoretical accuracy of attitude angle from the obtained accuracy of landmarks, and propose a way to improve the accuracy of landmarks. The model attitude angle acquisition method is verified by experimental measurements of a wind-wave tank model with a 1:50 Spar-type floating wind turbine model, the results show that the point measurement accuracy can reach sub-millimeter level, and the difference between the attitude angle measurement value and the theoretical value is less than 2.34′, which meets the testing requirements. The method proposed can be further extended to the attitude angle measurement in various wind tunnel pool tests.
在浮式海上风力机风箱模型试验中,模型姿态角的获取是关键目标之一。由于风浪罐所处的环境,传统接触式传感器不可避免地存在安装困难、模型重量增加、采样频率受限、数据采集不完整等缺点。本文提出了一种基于视觉的视频采集技术,主要包括:(1)构建高速视频采集网络;(2)对序列图像上的圆形地标进行识别、跟踪和匹配,利用空间点重建算法获得空间点的三维坐标;(3)提出了一种计算序列图像中模型姿态角的方法,以获得模型在每一帧中的位姿;(4)从得到的地标精度推导出姿态角的理论精度,并提出了提高地标精度的方法。通过对风浪箱模型与1:50桅杆式浮式风力机模型的实验测量验证了模型姿态角采集方法,结果表明,点测量精度可达到亚毫米级,姿态角测量值与理论值的差值小于2.34′,满足试验要求。该方法可进一步推广到各种风洞池试验的姿态角测量中。
{"title":"High-speed Video Measurement Method of Attitude Angle Applied to Floating Offshore Wind Turbine Model Wind-Wave Tank Test","authors":"Qing Zhong, Jiahao Liu, Peng Chen","doi":"10.1145/3581807.3581859","DOIUrl":"https://doi.org/10.1145/3581807.3581859","url":null,"abstract":"During the wind-wave tank model test of floating offshore wind turbine (FOWT), acquiring the model attitude angles is one of the key objectives. Due to the environment of the wind-wave tank, there are inevitable shortcomings of traditional contact sensors, such as difficult installation, increased model weight, limited sampling frequency, and incomplete data acquisition. In this paper, a vision-based technology is presented, which mainly consists of (1) construct a high-speed video acquisition network; (2) identify, track and match the circular landmarks on the sequence image, and obtain three-dimensional coordinates of the spatial points using the spatial point reconstruction algorithm; (3) propose a method for computing model attitude angles in sequence images to obtain the pose of the model in each frame; (4) deduce the theoretical accuracy of attitude angle from the obtained accuracy of landmarks, and propose a way to improve the accuracy of landmarks. The model attitude angle acquisition method is verified by experimental measurements of a wind-wave tank model with a 1:50 Spar-type floating wind turbine model, the results show that the point measurement accuracy can reach sub-millimeter level, and the difference between the attitude angle measurement value and the theoretical value is less than 2.34′, which meets the testing requirements. The method proposed can be further extended to the attitude angle measurement in various wind tunnel pool tests.","PeriodicalId":292813,"journal":{"name":"Proceedings of the 2022 11th International Conference on Computing and Pattern Recognition","volume":"537 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127058782","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Aneurysm Localization Algorithm Based on Faster R-CNN Network for Cerebral Small Vessels 基于更快R-CNN网络的脑小血管动脉瘤定位算法
Yuan Meng, Xinfeng Zhang, Xiaomin Liu, Xiangshen Li, Tianyu Zhu, Xiaoxia Chang, Jinhang Chen, Xiangyu Chen
The use of artificial intelligence algorithm to determine whether the lesion has cerebral aneurysm, especially small aneurysms, is still not completely solved. In this paper, the Faster R-CNN network was used as the localization network, and the model was trained by adjusting the network parameters, and the appropriate feature extraction network and classification network were selected to finally solve the localization problem of small aneurysms. Compared with most 3D methods, this method had the characteristics of shorter training cycle and faster image recognition. The experimental results show that the algorithm has a high accuracy in discriminating whether the lesion has cerebral aneurysm, but the false positive phenomenon may occur in the identification of single image localization. Finally, the paper discusses the experimental results and puts forward some conjecture ideas to solve the problem.
利用人工智能算法判断病变部位是否存在脑动脉瘤,尤其是小动脉瘤,目前仍未完全解决。本文采用Faster R-CNN网络作为定位网络,通过调整网络参数对模型进行训练,选择合适的特征提取网络和分类网络,最终解决小动脉瘤的定位问题。与大多数三维方法相比,该方法具有训练周期短、图像识别速度快的特点。实验结果表明,该算法在判断病灶是否存在脑动脉瘤方面具有较高的准确率,但在单幅图像定位的识别中可能出现假阳性现象。最后,对实验结果进行了讨论,并提出了一些解决问题的猜想。
{"title":"An Aneurysm Localization Algorithm Based on Faster R-CNN Network for Cerebral Small Vessels","authors":"Yuan Meng, Xinfeng Zhang, Xiaomin Liu, Xiangshen Li, Tianyu Zhu, Xiaoxia Chang, Jinhang Chen, Xiangyu Chen","doi":"10.1145/3581807.3581889","DOIUrl":"https://doi.org/10.1145/3581807.3581889","url":null,"abstract":"The use of artificial intelligence algorithm to determine whether the lesion has cerebral aneurysm, especially small aneurysms, is still not completely solved. In this paper, the Faster R-CNN network was used as the localization network, and the model was trained by adjusting the network parameters, and the appropriate feature extraction network and classification network were selected to finally solve the localization problem of small aneurysms. Compared with most 3D methods, this method had the characteristics of shorter training cycle and faster image recognition. The experimental results show that the algorithm has a high accuracy in discriminating whether the lesion has cerebral aneurysm, but the false positive phenomenon may occur in the identification of single image localization. Finally, the paper discusses the experimental results and puts forward some conjecture ideas to solve the problem.","PeriodicalId":292813,"journal":{"name":"Proceedings of the 2022 11th International Conference on Computing and Pattern Recognition","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133545298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Object Detection Algorithm Based on Coordinate Attention and Context Feature Enhancement 基于坐标关注和上下文特征增强的目标检测算法
Lingzhi Liu, Baohua Qiang, Yuan-yuan Wang, Xianyi Yang, Jubo Tian, S. Zhang
In recent years, object detection has been widely used in various fields such as face detection, remote sensing image detection and pedestrian detection. Due to the complex environment in the actual scene, we need to fully obtain the feature information in the image to improve the accuracy of object detection. This paper proposes an object detection algorithm based on coordinate attention and contextual feature enhancement. We design a multi-scale attention feature pyramid network, which first uses multi-branch atrous convolution to capture multi-scale context information, and then fuses the coordinate attention mechanism to embed location information into channel attention, and finally uses a bidirectional feature pyramid structure to effectively fuse high-level features and low-level features. We also adopt the GIoU loss function to further improve the accuracy of object detection. The experimental results show that the proposed method has certain advantages compared with other detection algorithms in the PASCAL VOC datasets.
近年来,物体检测在人脸检测、遥感图像检测、行人检测等各个领域得到了广泛的应用。由于实际场景环境复杂,我们需要充分获取图像中的特征信息,以提高目标检测的精度。提出了一种基于坐标关注和上下文特征增强的目标检测算法。设计了一个多尺度关注特征金字塔网络,首先利用多分支亚属性卷积捕获多尺度上下文信息,然后融合坐标关注机制将位置信息嵌入到通道关注中,最后利用双向特征金字塔结构有效融合高阶特征和低阶特征。我们还采用了GIoU损失函数,进一步提高了目标检测的精度。实验结果表明,在PASCAL VOC数据集上,与其他检测算法相比,该方法具有一定的优势。
{"title":"Object Detection Algorithm Based on Coordinate Attention and Context Feature Enhancement","authors":"Lingzhi Liu, Baohua Qiang, Yuan-yuan Wang, Xianyi Yang, Jubo Tian, S. Zhang","doi":"10.1145/3581807.3581821","DOIUrl":"https://doi.org/10.1145/3581807.3581821","url":null,"abstract":"In recent years, object detection has been widely used in various fields such as face detection, remote sensing image detection and pedestrian detection. Due to the complex environment in the actual scene, we need to fully obtain the feature information in the image to improve the accuracy of object detection. This paper proposes an object detection algorithm based on coordinate attention and contextual feature enhancement. We design a multi-scale attention feature pyramid network, which first uses multi-branch atrous convolution to capture multi-scale context information, and then fuses the coordinate attention mechanism to embed location information into channel attention, and finally uses a bidirectional feature pyramid structure to effectively fuse high-level features and low-level features. We also adopt the GIoU loss function to further improve the accuracy of object detection. The experimental results show that the proposed method has certain advantages compared with other detection algorithms in the PASCAL VOC datasets.","PeriodicalId":292813,"journal":{"name":"Proceedings of the 2022 11th International Conference on Computing and Pattern Recognition","volume":"94 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124244885","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
DOA Estimation of Multiple Sources based on the Angle Distribution of Time-frequency Points in Single-source Zone 基于时频点角度分布的多源DOA估计
Liang Tao, Mao-shen Jia, Lu Li
Direction-of-arrival (DOA) estimation method based on single-source zone (SSZ) detection, using the sparsity of speech signal, which transforms the multiple sources localization into single source localization. However, there are many time-frequency (TF) points whose direction information are far away from the true DOA in the detected SSZ, these points may disturb the localization performance. Aiming this issue, a DOA estimation of multiple sources based on the angle distribution of TF points is proposed in this paper. Firstly, the SSZs are detected through the recorded signal of sound field microphone. Secondly, the optimized single-source zone (OSSZ) can be acquired by removing the outliers based on the angle distribution of the TF points in the detected SSZ. Thirdly, DOA histogram can be obtained using the TF points in OSSZ, then the envelop of the DOA histogram is gained by kernel density estimation. Finally, peak search is adopted to obtain the DOA estimates and number of sources. The experiment results show that the proposed method can achieve better localization performance than SSZ-based method under medium and high reverberation conditions.
基于单源区域(SSZ)检测的DOA估计方法,利用语音信号的稀疏性,将多源定位转化为单源定位。然而,在被检测的SSZ中存在许多时频点,它们的方向信息与真实DOA相差甚远,这些点可能会干扰定位性能。针对这一问题,本文提出了一种基于TF点角度分布的多源DOA估计方法。首先,通过声场传声器记录的信号来检测ssz。其次,根据检测到的单源区域内TF点的角度分布,去除异常值,得到优化的单源区域(OSSZ);第三,利用OSSZ中的TF点得到DOA直方图,然后通过核密度估计得到DOA直方图的包络。最后,采用峰值搜索方法,得到信号的DOA估计值和信源数量。实验结果表明,在中、高混响条件下,该方法比基于ssz的方法具有更好的定位性能。
{"title":"DOA Estimation of Multiple Sources based on the Angle Distribution of Time-frequency Points in Single-source Zone","authors":"Liang Tao, Mao-shen Jia, Lu Li","doi":"10.1145/3581807.3581861","DOIUrl":"https://doi.org/10.1145/3581807.3581861","url":null,"abstract":"Direction-of-arrival (DOA) estimation method based on single-source zone (SSZ) detection, using the sparsity of speech signal, which transforms the multiple sources localization into single source localization. However, there are many time-frequency (TF) points whose direction information are far away from the true DOA in the detected SSZ, these points may disturb the localization performance. Aiming this issue, a DOA estimation of multiple sources based on the angle distribution of TF points is proposed in this paper. Firstly, the SSZs are detected through the recorded signal of sound field microphone. Secondly, the optimized single-source zone (OSSZ) can be acquired by removing the outliers based on the angle distribution of the TF points in the detected SSZ. Thirdly, DOA histogram can be obtained using the TF points in OSSZ, then the envelop of the DOA histogram is gained by kernel density estimation. Finally, peak search is adopted to obtain the DOA estimates and number of sources. The experiment results show that the proposed method can achieve better localization performance than SSZ-based method under medium and high reverberation conditions.","PeriodicalId":292813,"journal":{"name":"Proceedings of the 2022 11th International Conference on Computing and Pattern Recognition","volume":"207 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123018714","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Design Method of SSVEP Stimulus Source based on Overlooking Map 基于俯瞰图的SSVEP激励源设计方法
Dong Wen, Mengmeng Jiang, Wenlong Jiao, Xianglong Wan, Xifa Lan, Yanhong Zhou
In the field of brain-computer interface, steady-state visual evoked potential (SSVEP) is widely used because of its stability. Although high-intensity stimulus has good accuracy, it can cause severe visual fatigue and even induce epilepsy in subjects. As well as paying attention to its accuracy, personnel should also pay attention to the comfort of the subject. In this paper, combined with the knowledge of spatial psychology, the overlooking map is proposed as the stimulus source to induce the SSVEP signal. 24 subjects participated in the comparison experiment between the overlooking map and the black and white stimuli, and the EEG signal was processed and analyzed online in real time through the CCA algorithm. Afterwards, the two stimuli were scored in terms of personal preference, comfort, and flickering sensation. The experimental results show that the performance of the overlooking map stimulus source is superior to that of the black and white stimulus, and it is more suitable to induce the SSVEP signal of the subjects. As an important aspect of SSVEP-based application and a necessary factor for commercial promotion, user experience provides a good theoretical and experimental research basis for it.
稳态视觉诱发电位(SSVEP)因其稳定性被广泛应用于脑机接口领域。高强度刺激虽然具有良好的准确性,但会引起受试者严重的视觉疲劳,甚至诱发癫痫。人员在注意其准确性的同时,还应注意拍摄对象的舒适度。本文结合空间心理学的相关知识,提出瞭望图作为诱发SSVEP信号的刺激源。24名被试参与了俯瞰图与黑白刺激的对比实验,并通过CCA算法对EEG信号进行在线实时处理和分析。之后,根据个人偏好、舒适度和闪烁感对这两种刺激进行评分。实验结果表明,俯视图刺激源的性能优于黑白刺激源,更适合诱导被试的SSVEP信号。用户体验作为ssvep应用的重要方面和商业推广的必要因素,为ssvep提供了良好的理论和实验研究基础。
{"title":"The Design Method of SSVEP Stimulus Source based on Overlooking Map","authors":"Dong Wen, Mengmeng Jiang, Wenlong Jiao, Xianglong Wan, Xifa Lan, Yanhong Zhou","doi":"10.1145/3581807.3581874","DOIUrl":"https://doi.org/10.1145/3581807.3581874","url":null,"abstract":"In the field of brain-computer interface, steady-state visual evoked potential (SSVEP) is widely used because of its stability. Although high-intensity stimulus has good accuracy, it can cause severe visual fatigue and even induce epilepsy in subjects. As well as paying attention to its accuracy, personnel should also pay attention to the comfort of the subject. In this paper, combined with the knowledge of spatial psychology, the overlooking map is proposed as the stimulus source to induce the SSVEP signal. 24 subjects participated in the comparison experiment between the overlooking map and the black and white stimuli, and the EEG signal was processed and analyzed online in real time through the CCA algorithm. Afterwards, the two stimuli were scored in terms of personal preference, comfort, and flickering sensation. The experimental results show that the performance of the overlooking map stimulus source is superior to that of the black and white stimulus, and it is more suitable to induce the SSVEP signal of the subjects. As an important aspect of SSVEP-based application and a necessary factor for commercial promotion, user experience provides a good theoretical and experimental research basis for it.","PeriodicalId":292813,"journal":{"name":"Proceedings of the 2022 11th International Conference on Computing and Pattern Recognition","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130550058","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust Deep Convolutional Neural Network inspired by the Primary Visual Cortex 受初级视觉皮层启发的鲁棒深度卷积神经网络
Zhanguo Dong, Ming Ke, Jiarong Wang, Lubin Wang, Gang Wang
Most of the current advanced object recognition deep convolutional neural networks (DCNNs) are vulnerable to attacks of adversarial perturbations. In comparison, the primate vision system can effectively suppress the inference of adversarial perturbations. Many studies have shown that the fusion of biological vision mechanisms and DCNNs is a promising way to improve model robustness. The primary visual cortex (V1) is a key brain region for visual information processing in the biological brain, containing various simple cell orientation selection receptive fields, which can specifically respond to low-level features. Therefore, we have developed an object classification DCNN model inspired by V1 orientation selection receptive fields. The V1-inspired model introduces V1 orientation selection receptive fields into DCNN through anisotropic Gaussian kernels, which can enrich the receptive fields of DCNN. In the white-box adversarial attack experiments on CIFAR-100 and Mini-ImageNet, the adversarial robustness of our model is 21.74% and 20.01% higher than that of the baseline DCNN, respectively. Compared with the SOAT VOneNet, the adversarial robustness of our model improves by 2.88% and 8.56%, respectively. It is worth pointing out that our method will not increase the parameter quantity of the baseline model, while the extra training cost is very little.
目前大多数先进的目标识别深度卷积神经网络(DCNNs)都容易受到对抗性扰动的攻击。相比之下,灵长类视觉系统可以有效地抑制对抗性扰动的推理。许多研究表明,生物视觉机制与DCNNs的融合是提高模型鲁棒性的一种很有前途的方法。初级视觉皮层(primary visual cortex, V1)是生物脑中处理视觉信息的关键脑区,包含多种简单的细胞定向选择感受野,对低水平特征具有特异性反应。因此,我们开发了一个基于V1方向选择接受野的目标分类DCNN模型。V1启发模型通过各向异性高斯核将V1取向选择感受场引入DCNN,丰富了DCNN的感受场。在CIFAR-100和Mini-ImageNet的白盒对抗攻击实验中,我们的模型的对抗鲁棒性分别比基线DCNN高21.74%和20.01%。与SOAT VOneNet相比,该模型的对抗鲁棒性分别提高了2.88%和8.56%。值得指出的是,我们的方法不会增加基线模型的参数数量,而额外的训练成本也很少。
{"title":"Robust Deep Convolutional Neural Network inspired by the Primary Visual Cortex","authors":"Zhanguo Dong, Ming Ke, Jiarong Wang, Lubin Wang, Gang Wang","doi":"10.1145/3581807.3581893","DOIUrl":"https://doi.org/10.1145/3581807.3581893","url":null,"abstract":"Most of the current advanced object recognition deep convolutional neural networks (DCNNs) are vulnerable to attacks of adversarial perturbations. In comparison, the primate vision system can effectively suppress the inference of adversarial perturbations. Many studies have shown that the fusion of biological vision mechanisms and DCNNs is a promising way to improve model robustness. The primary visual cortex (V1) is a key brain region for visual information processing in the biological brain, containing various simple cell orientation selection receptive fields, which can specifically respond to low-level features. Therefore, we have developed an object classification DCNN model inspired by V1 orientation selection receptive fields. The V1-inspired model introduces V1 orientation selection receptive fields into DCNN through anisotropic Gaussian kernels, which can enrich the receptive fields of DCNN. In the white-box adversarial attack experiments on CIFAR-100 and Mini-ImageNet, the adversarial robustness of our model is 21.74% and 20.01% higher than that of the baseline DCNN, respectively. Compared with the SOAT VOneNet, the adversarial robustness of our model improves by 2.88% and 8.56%, respectively. It is worth pointing out that our method will not increase the parameter quantity of the baseline model, while the extra training cost is very little.","PeriodicalId":292813,"journal":{"name":"Proceedings of the 2022 11th International Conference on Computing and Pattern Recognition","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127371863","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Heart Rate Detection Using Motion Compensation with Multiple ROIs 基于多roi运动补偿的心率检测
Jie Huang, Xuanheng Rao, Weichuan Zhang, Jingze Song, Xiao Sun
Remote photoplethysmography (rPPG) has the ability to make use of image frame sequences including human faces collected by cameras for measuring heart rate (HR) without any contact. This method generates a time series signal based on the RGB spatial average of the selected region of interest (ROI) to estimate physiological signals such as HR. It is worth to note that the motion artifact produced by the subject’s face shaking is equivalent to adding considerable noise to the signal which will greatly affect the accuracy of the measurement. In this paper, a novel anti-interference multi-ROI analysis (AMA) approach is proposed which effectively utilizes the local information with multiple ROIs, the Euler angle information of the subject’s head, and the interpolation resampling technique of the video for suppressing the influence of face shaking on the accuracy of non-contact heart rate measurement. The proposed method is evaluated on the UBFC-RPPG and PURE datasets, and the experimental results demonstrate that the proposed methods are superior to many state-of-the-art methods.
远程光电容积脉搏描记(rPPG)能够利用相机收集的包括人脸在内的图像帧序列,在没有任何接触的情况下测量心率(HR)。该方法基于所选感兴趣区域(ROI)的RGB空间平均值生成时间序列信号,用于估计HR等生理信号。值得注意的是,受试者面部抖动产生的运动伪影相当于在信号中增加了相当大的噪声,这将极大地影响测量的准确性。本文提出了一种新的抗干扰多roi分析方法,该方法有效地利用了具有多个roi的局部信息、被测者头部的欧拉角信息以及视频的插值重采样技术来抑制面部抖动对非接触式心率测量精度的影响。在UBFC-RPPG和PURE数据集上对所提方法进行了评估,实验结果表明,所提方法优于许多最先进的方法。
{"title":"Heart Rate Detection Using Motion Compensation with Multiple ROIs","authors":"Jie Huang, Xuanheng Rao, Weichuan Zhang, Jingze Song, Xiao Sun","doi":"10.1145/3581807.3581870","DOIUrl":"https://doi.org/10.1145/3581807.3581870","url":null,"abstract":"Remote photoplethysmography (rPPG) has the ability to make use of image frame sequences including human faces collected by cameras for measuring heart rate (HR) without any contact. This method generates a time series signal based on the RGB spatial average of the selected region of interest (ROI) to estimate physiological signals such as HR. It is worth to note that the motion artifact produced by the subject’s face shaking is equivalent to adding considerable noise to the signal which will greatly affect the accuracy of the measurement. In this paper, a novel anti-interference multi-ROI analysis (AMA) approach is proposed which effectively utilizes the local information with multiple ROIs, the Euler angle information of the subject’s head, and the interpolation resampling technique of the video for suppressing the influence of face shaking on the accuracy of non-contact heart rate measurement. The proposed method is evaluated on the UBFC-RPPG and PURE datasets, and the experimental results demonstrate that the proposed methods are superior to many state-of-the-art methods.","PeriodicalId":292813,"journal":{"name":"Proceedings of the 2022 11th International Conference on Computing and Pattern Recognition","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126886212","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Portrait Interpretation and a Benchmark 《肖像解读与基准
Yixuan Fan, Zhaopeng Dou, Yali Li, Shengjin Wang
We propose a task we name Portrait Interpretation and construct a dataset named Portrait250K for it. Current researches on portraits such as human attribute recognition and person re-identification have achieved many successes, but generally, they: 1) may lack mining the interrelationship between various tasks and the possible benefits it may bring; 2) design deep models specifically for each task, which is inefficient; 3) may be unable to cope with the needs of a unified model and comprehensive perception in actual scenes. In this paper, the proposed portrait interpretation recognizes the perception of humans from a new systematic perspective. We divide the perception of portraits into three aspects, namely Appearance, Posture, and Emotion, and design corresponding sub-tasks for each aspect. Based on the framework of multi-task learning, portrait interpretation requires a comprehensive description of static attributes and dynamic states of portraits. To invigorate research on this new task, we construct a new dataset that contains 250,000 images labeled with identity, gender, age, physique, height, expression, and posture of the whole body and arms. Our dataset is collected from 51 movies, hence covering extensive diversity. Furthermore, we focus on representation learning for portrait interpretation and propose a baseline that reflects our systematic perspective. We also propose an appropriate metric for this task. Our experimental results demonstrate that combining the tasks related to portrait interpretation can yield benefits. Code and dataset will be made public.
我们提出了一个名为Portrait Interpretation的任务,并为此构建了一个名为porttrait250k的数据集。目前关于人像识别的研究,如人的属性识别和人的再识别等,取得了不少成果,但总体上存在以下问题:1)缺乏挖掘各种任务之间的相互关系及其可能带来的利益;2)针对每个任务专门设计深度模型,效率低下;3)在实际场景中可能无法应对统一模型和全面感知的需求。在本文中,提出的肖像解释从一个新的系统角度认识了人类的感知。我们将人像感知分为外貌、姿态和情感三个方面,并针对每个方面设计相应的子任务。肖像解读基于多任务学习的框架,需要对肖像的静态属性和动态状态进行综合描述。为了激发对这项新任务的研究,我们构建了一个新的数据集,其中包含25万张标记为身份、性别、年龄、体格、身高、表情和全身和手臂姿势的图像。我们的数据集来自51部电影,因此涵盖了广泛的多样性。此外,我们将重点放在肖像解释的表征学习上,并提出了一个反映我们系统视角的基线。我们还为这项任务提出了一个适当的度量。我们的实验结果表明,结合与肖像解释相关的任务可以产生好处。代码和数据集将公开。
{"title":"Portrait Interpretation and a Benchmark","authors":"Yixuan Fan, Zhaopeng Dou, Yali Li, Shengjin Wang","doi":"10.1145/3581807.3581838","DOIUrl":"https://doi.org/10.1145/3581807.3581838","url":null,"abstract":"We propose a task we name Portrait Interpretation and construct a dataset named Portrait250K for it. Current researches on portraits such as human attribute recognition and person re-identification have achieved many successes, but generally, they: 1) may lack mining the interrelationship between various tasks and the possible benefits it may bring; 2) design deep models specifically for each task, which is inefficient; 3) may be unable to cope with the needs of a unified model and comprehensive perception in actual scenes. In this paper, the proposed portrait interpretation recognizes the perception of humans from a new systematic perspective. We divide the perception of portraits into three aspects, namely Appearance, Posture, and Emotion, and design corresponding sub-tasks for each aspect. Based on the framework of multi-task learning, portrait interpretation requires a comprehensive description of static attributes and dynamic states of portraits. To invigorate research on this new task, we construct a new dataset that contains 250,000 images labeled with identity, gender, age, physique, height, expression, and posture of the whole body and arms. Our dataset is collected from 51 movies, hence covering extensive diversity. Furthermore, we focus on representation learning for portrait interpretation and propose a baseline that reflects our systematic perspective. We also propose an appropriate metric for this task. Our experimental results demonstrate that combining the tasks related to portrait interpretation can yield benefits. Code and dataset will be made public.","PeriodicalId":292813,"journal":{"name":"Proceedings of the 2022 11th International Conference on Computing and Pattern Recognition","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-07-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127135701","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Proceedings of the 2022 11th International Conference on Computing and Pattern Recognition 2022年第11届计算与模式识别国际会议论文集
{"title":"Proceedings of the 2022 11th International Conference on Computing and Pattern Recognition","authors":"","doi":"10.1145/3581807","DOIUrl":"https://doi.org/10.1145/3581807","url":null,"abstract":"","PeriodicalId":292813,"journal":{"name":"Proceedings of the 2022 11th International Conference on Computing and Pattern Recognition","volume":"61 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125166395","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the 2022 11th International Conference on Computing and Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1