首页 > 最新文献

2022 International Joint Conference on Neural Networks (IJCNN)最新文献

英文 中文
Transferring multiple text styles using CycleGAN with supervised style latent space 使用带有监督样式潜在空间的CycleGAN转移多个文本样式
Pub Date : 2022-07-18 DOI: 10.1109/IJCNN55064.2022.9892978
Lorenzo Puppi Vecchi, E. C. F. Maffezzolli, E. Paraiso
Text style transfer is a relevant task, contributing to theoretical and practical advancement in several areas, especially when working with non-parallel data. The concept behind non-parallel style transfer is to change a specific dimension of the sentence while retaining the overall context. Previous work used adversarial learning to perform such a task. Although it was not initially created to work with textual data, it proved very effective. Most of the previous work has focused on developing algorithms capable of transferring between binary styles, with limited generalization capabilities and limited applications. This work proposes a framework capable of working with multiple styles and improving content retention (BLEU) after a transfer. The proposed framework combines supervised learning of latent spaces and their separation within the architecture. The results suggest that the proposed framework improves content retention in multi-style scenarios while maintaining accuracy comparable to state-of-the-art.
文本样式转移是一项相关的任务,在几个领域,特别是在处理非并行数据时,有助于理论和实践的进步。非平行文体迁移背后的概念是在保留整体语境的同时改变句子的特定维度。以前的工作使用对抗性学习来完成这样的任务。虽然它最初不是为处理文本数据而创建的,但事实证明它非常有效。以前的大部分工作都集中在开发能够在二进制样式之间转换的算法上,泛化能力有限,应用也有限。这项工作提出了一个框架能够工作与多种风格和提高内容保留(BLEU)后转移。提出的框架结合了潜在空间的监督学习和它们在建筑中的分离。结果表明,所提出的框架提高了多风格场景中的内容保留,同时保持了与最新技术相当的准确性。
{"title":"Transferring multiple text styles using CycleGAN with supervised style latent space","authors":"Lorenzo Puppi Vecchi, E. C. F. Maffezzolli, E. Paraiso","doi":"10.1109/IJCNN55064.2022.9892978","DOIUrl":"https://doi.org/10.1109/IJCNN55064.2022.9892978","url":null,"abstract":"Text style transfer is a relevant task, contributing to theoretical and practical advancement in several areas, especially when working with non-parallel data. The concept behind non-parallel style transfer is to change a specific dimension of the sentence while retaining the overall context. Previous work used adversarial learning to perform such a task. Although it was not initially created to work with textual data, it proved very effective. Most of the previous work has focused on developing algorithms capable of transferring between binary styles, with limited generalization capabilities and limited applications. This work proposes a framework capable of working with multiple styles and improving content retention (BLEU) after a transfer. The proposed framework combines supervised learning of latent spaces and their separation within the architecture. The results suggest that the proposed framework improves content retention in multi-style scenarios while maintaining accuracy comparable to state-of-the-art.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114755125","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient Meta-Learning for Continual Learning with Taylor Expansion Approximation 基于泰勒展开近似的持续学习的有效元学习
Pub Date : 2022-07-18 DOI: 10.1109/IJCNN55064.2022.9892669
Xiaohan Zou, Tong Lin
Continual learning aims to alleviate catastrophic forgetting when handling consecutive tasks under non-stationary distributions. Gradient-based meta-learning algorithms have shown the capability to implicitly solve the transfer-interference trade-off problem between different examples. However, they still suffer from the catastrophic forgetting problem in the setting of continual learning, since the past data of previous tasks are no longer available. In this work, we propose a novel efficient meta-learning algorithm for solving the online continual learning problem, where the regularization terms and learning rates are adapted to the Taylor approximation of the parameter's importance to mitigate forgetting. The proposed method expresses the gradient of the meta-loss in closed-form and thus avoid computing second-order derivative which is computationally inhibitable. We also use Proximal Gradient Descent to further improve computational efficiency and accuracy. Experiments on diverse benchmarks show that our method achieves better or on-par performance and much higher efficiency compared to the state-of-the-art approaches.
持续学习旨在减轻在非平稳分布下处理连续任务时的灾难性遗忘。基于梯度的元学习算法已经显示出隐式解决不同示例之间迁移-干扰权衡问题的能力。然而,在持续学习的情况下,他们仍然存在灾难性的遗忘问题,因为以前任务的过去数据不再可用。在这项工作中,我们提出了一种新的有效的元学习算法来解决在线持续学习问题,其中正则化项和学习率适应参数重要性的泰勒近似,以减轻遗忘。该方法以封闭形式表示元损失的梯度,从而避免了二阶导数的计算抑制。为了进一步提高计算效率和精度,我们还使用了近端梯度下降。在不同基准上的实验表明,与最先进的方法相比,我们的方法实现了更好或同等的性能和更高的效率。
{"title":"Efficient Meta-Learning for Continual Learning with Taylor Expansion Approximation","authors":"Xiaohan Zou, Tong Lin","doi":"10.1109/IJCNN55064.2022.9892669","DOIUrl":"https://doi.org/10.1109/IJCNN55064.2022.9892669","url":null,"abstract":"Continual learning aims to alleviate catastrophic forgetting when handling consecutive tasks under non-stationary distributions. Gradient-based meta-learning algorithms have shown the capability to implicitly solve the transfer-interference trade-off problem between different examples. However, they still suffer from the catastrophic forgetting problem in the setting of continual learning, since the past data of previous tasks are no longer available. In this work, we propose a novel efficient meta-learning algorithm for solving the online continual learning problem, where the regularization terms and learning rates are adapted to the Taylor approximation of the parameter's importance to mitigate forgetting. The proposed method expresses the gradient of the meta-loss in closed-form and thus avoid computing second-order derivative which is computationally inhibitable. We also use Proximal Gradient Descent to further improve computational efficiency and accuracy. Experiments on diverse benchmarks show that our method achieves better or on-par performance and much higher efficiency compared to the state-of-the-art approaches.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124301889","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Siamese Network Tracker by Attention Module and Relation Detector Module 由关注模块和关系检测模块组成的暹罗网络跟踪器
Pub Date : 2022-07-18 DOI: 10.1109/IJCNN55064.2022.9892067
Xiaohan Liu, Aimin Li, Deqi Liu, Dexu Yao, Mengfan Cheng
In recent years, object tracking techniques based on Siamese networks have shown excellent tracking performance. However, in the tracking process, there will be many similar objects, and it is easy to track the wrong object due to the weak discriminative ability of the network. At the same time, the classification and regression of SiamRPN ++ are usually optimized independently, which will cause a mismatch problem, that is, the location with the highest classification confidence is not necessarily the object. To address these problems, we proposed a Siamese network tracker by attention module and relation detector module (SiamAR). First, we introduce a multi-scale attention mechanism in SiamRPN++ to capture information at different scales, and fuse spatial attention and channel attention to improving the ability to learn feature information. Not only different receptive fields are obtained, but also useful features are selectively focused and less useful features are suppressed. In order not to affect the computational efficiency, the method of grouping parallel computing is used. Secondly, we add a relation detector module to our tracker to filter out distractors from the background and distinguish the object in the cluttered background. Experiment results show that our algorithm out-performs several well-known tracking algorithms in terms of tracking accuracy and robustness.
近年来,基于Siamese网络的目标跟踪技术表现出了良好的跟踪性能。然而,在跟踪过程中,会有许多相似的对象,由于网络的判别能力较弱,很容易跟踪到错误的对象。同时,siamrpn++的分类和回归通常是独立优化的,这会造成不匹配问题,即分类置信度最高的位置不一定是目标。为了解决这些问题,我们提出了一种由关注模块和关系检测模块(SiamAR)组成的Siamese网络跟踪器。首先,在siamrpn++中引入多尺度注意机制,捕获不同尺度的信息,融合空间注意和通道注意,提高特征信息的学习能力;不仅获得了不同的感受野,而且有用的特征被选择性地聚焦,不太有用的特征被抑制。为了不影响计算效率,采用分组并行计算的方法。其次,我们在跟踪器中加入关系检测模块,过滤掉背景中的干扰物,并在杂乱的背景中区分目标。实验结果表明,该算法在跟踪精度和鲁棒性方面优于几种已知的跟踪算法。
{"title":"Siamese Network Tracker by Attention Module and Relation Detector Module","authors":"Xiaohan Liu, Aimin Li, Deqi Liu, Dexu Yao, Mengfan Cheng","doi":"10.1109/IJCNN55064.2022.9892067","DOIUrl":"https://doi.org/10.1109/IJCNN55064.2022.9892067","url":null,"abstract":"In recent years, object tracking techniques based on Siamese networks have shown excellent tracking performance. However, in the tracking process, there will be many similar objects, and it is easy to track the wrong object due to the weak discriminative ability of the network. At the same time, the classification and regression of SiamRPN ++ are usually optimized independently, which will cause a mismatch problem, that is, the location with the highest classification confidence is not necessarily the object. To address these problems, we proposed a Siamese network tracker by attention module and relation detector module (SiamAR). First, we introduce a multi-scale attention mechanism in SiamRPN++ to capture information at different scales, and fuse spatial attention and channel attention to improving the ability to learn feature information. Not only different receptive fields are obtained, but also useful features are selectively focused and less useful features are suppressed. In order not to affect the computational efficiency, the method of grouping parallel computing is used. Secondly, we add a relation detector module to our tracker to filter out distractors from the background and distinguish the object in the cluttered background. Experiment results show that our algorithm out-performs several well-known tracking algorithms in terms of tracking accuracy and robustness.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127886975","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A One-Shot Reparameterization Method for Reducing the Loss of Tile Pruning on DNNs 一种减少dnn剪枝损失的单次重参数化方法
Pub Date : 2022-07-18 DOI: 10.1109/IJCNN55064.2022.9889789
Yancheng Li, Qingzhong Ai, Fumihiko Ino
Recently, tile pruning has been widely studied to accelerate the inference of deep neural networks (DNNs). However, we found that the loss due to tile pruning, which can eliminate important elements together with unimportant elements, is large on trained DNNs. In this study, we propose a one-shot reparameterization method, called TileTrans, to reduce the loss of tile pruning. Specifically, we repermute the rows or columns of the weight matrix such that the model architecture can be kept unchanged after reparameterization. This repermutation realizes the reparameterization of the DNN model without any retraining. The proposed reparameterization method combines important elements into the same tile; thus, preserving the important elements after the tile pruning. Furthermore, TileTrans can be seamlessly integrated into existing tile pruning methods because it is a pre-processing method executed before pruning, which is orthogonal to most existing methods. The experimental results demonstrate that our method is essential in reducing the loss of tile pruning on DNNs. Specifically, the accuracy is improved by up to 17% for AlexNet while 5% for ResNet-34, where both models are pre-trained on ImageNet.
近年来,为了加速深度神经网络(dnn)的推理,瓦片修剪得到了广泛的研究。然而,我们发现,在训练好的dnn上,由于瓷砖修剪可以去除重要元素和不重要元素而造成的损失很大。在这项研究中,我们提出了一种一次性重新参数化方法,称为TileTrans,以减少瓷砖修剪的损失。具体来说,我们重新调整权重矩阵的行或列,使模型架构在重新参数化后保持不变。这种再突变在不进行再训练的情况下实现了DNN模型的再参数化。提出的重新参数化方法将重要元素合并到同一块图中;因此,在瓷砖修剪后保留了重要的元素。此外,TileTrans可以无缝集成到现有的瓷砖修剪方法中,因为它是在修剪之前执行的预处理方法,与大多数现有方法正交。实验结果表明,我们的方法对于减少dnn上的剪枝损失是必不可少的。具体来说,AlexNet的准确率提高了17%,而ResNet-34的准确率提高了5%,这两个模型都是在ImageNet上进行预训练的。
{"title":"A One-Shot Reparameterization Method for Reducing the Loss of Tile Pruning on DNNs","authors":"Yancheng Li, Qingzhong Ai, Fumihiko Ino","doi":"10.1109/IJCNN55064.2022.9889789","DOIUrl":"https://doi.org/10.1109/IJCNN55064.2022.9889789","url":null,"abstract":"Recently, tile pruning has been widely studied to accelerate the inference of deep neural networks (DNNs). However, we found that the loss due to tile pruning, which can eliminate important elements together with unimportant elements, is large on trained DNNs. In this study, we propose a one-shot reparameterization method, called TileTrans, to reduce the loss of tile pruning. Specifically, we repermute the rows or columns of the weight matrix such that the model architecture can be kept unchanged after reparameterization. This repermutation realizes the reparameterization of the DNN model without any retraining. The proposed reparameterization method combines important elements into the same tile; thus, preserving the important elements after the tile pruning. Furthermore, TileTrans can be seamlessly integrated into existing tile pruning methods because it is a pre-processing method executed before pruning, which is orthogonal to most existing methods. The experimental results demonstrate that our method is essential in reducing the loss of tile pruning on DNNs. Specifically, the accuracy is improved by up to 17% for AlexNet while 5% for ResNet-34, where both models are pre-trained on ImageNet.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126544379","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MSR-DARTS: Minimum Stable Rank of Differentiable Architecture Search MSR-DARTS:可微结构搜索的最小稳定秩
Pub Date : 2022-07-18 DOI: 10.1109/IJCNN55064.2022.9892751
Kengo Machida, K. Uto, K. Shinoda, Taiji Suzuki
In neural architecture search (NAS), differentiable architecture search (DARTS) has recently attracted much attention due to its high efficiency. However, this method finds a model with the weights converging faster than the others, and such a model with fastest convergence often leads to overfitting. Accordingly, the resulting model cannot always be well-generalized. To overcome this problem, we propose a method called minimum stable rank DARTS (MSR-DARTS), for finding a model with the best generalization error by replacing architecture optimization with the selection process using the minimum stable rank criterion. Specifically, a convolution operator is represented by a matrix, and MSR-DARTS selects the one with the smallest stable rank. We evaluated MSR-DARTS on CIFAR-10 and ImageNet datasets. It achieves an error rate of 2.54% with 4.0M parameters within 0.3 GPU-days on CIFAR-10, and a top-1 error rate of 23.9% on ImageNet.
在神经结构搜索(NAS)中,可微结构搜索(DARTS)因其高效率而受到广泛关注。然而,该方法寻找的是权重收敛速度快于其他模型的模型,而这种收敛速度快的模型往往会导致过拟合。因此,得到的模型不能总是很好地一般化。为了克服这一问题,我们提出了一种称为最小稳定秩DARTS (MSR-DARTS)的方法,通过使用最小稳定秩准则的选择过程取代架构优化来寻找具有最佳泛化误差的模型。具体来说,卷积算子用矩阵表示,MSR-DARTS选择稳定秩最小的卷积算子。我们在CIFAR-10和ImageNet数据集上评估了msr - dart。在CIFAR-10上,在0.3个gpu天内,在4.0M参数下,错误率为2.54%,在ImageNet上,错误率为23.9%。
{"title":"MSR-DARTS: Minimum Stable Rank of Differentiable Architecture Search","authors":"Kengo Machida, K. Uto, K. Shinoda, Taiji Suzuki","doi":"10.1109/IJCNN55064.2022.9892751","DOIUrl":"https://doi.org/10.1109/IJCNN55064.2022.9892751","url":null,"abstract":"In neural architecture search (NAS), differentiable architecture search (DARTS) has recently attracted much attention due to its high efficiency. However, this method finds a model with the weights converging faster than the others, and such a model with fastest convergence often leads to overfitting. Accordingly, the resulting model cannot always be well-generalized. To overcome this problem, we propose a method called minimum stable rank DARTS (MSR-DARTS), for finding a model with the best generalization error by replacing architecture optimization with the selection process using the minimum stable rank criterion. Specifically, a convolution operator is represented by a matrix, and MSR-DARTS selects the one with the smallest stable rank. We evaluated MSR-DARTS on CIFAR-10 and ImageNet datasets. It achieves an error rate of 2.54% with 4.0M parameters within 0.3 GPU-days on CIFAR-10, and a top-1 error rate of 23.9% on ImageNet.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":"33 5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125712295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Few-shot Graph Classification with Contrastive Loss and Meta-classifier 基于对比损失和元分类器的少射图分类
Pub Date : 2022-07-18 DOI: 10.1109/IJCNN55064.2022.9892886
Chao Wei, Zhidong Deng
Few-shot graph-level classification based on graph neural networks is critical in many tasks including drug and material discovery. We present a novel graph contrastive relation network (GCRNet) by introducing a practical yet straightforward graph meta-baseline with contrastive loss to gain robust representation and meta-classifier to realize more suitable similarity metric, which is more adaptive for graph few-shot problems. Experimental results demonstrate that the proposed method achieves 8%-12% in 5-shot, 5%-8% in 10 shot, and 1%-5% in 20-shot improvements, respectively, compared to the existing state-of-the-art methods.
基于图神经网络的少镜头图级分类在药物和材料发现等许多任务中至关重要。本文提出了一种新型的图对比关系网络(GCRNet),通过引入一种实用而直观的具有对比损失的图元基线来获得鲁棒表示,并引入元分类器来实现更合适的相似度度量,从而更适应图少射问题。实验结果表明,与现有方法相比,该方法的5次、10次和20次分别提高了8% ~ 12%、5% ~ 8%和1% ~ 5%。
{"title":"Few-shot Graph Classification with Contrastive Loss and Meta-classifier","authors":"Chao Wei, Zhidong Deng","doi":"10.1109/IJCNN55064.2022.9892886","DOIUrl":"https://doi.org/10.1109/IJCNN55064.2022.9892886","url":null,"abstract":"Few-shot graph-level classification based on graph neural networks is critical in many tasks including drug and material discovery. We present a novel graph contrastive relation network (GCRNet) by introducing a practical yet straightforward graph meta-baseline with contrastive loss to gain robust representation and meta-classifier to realize more suitable similarity metric, which is more adaptive for graph few-shot problems. Experimental results demonstrate that the proposed method achieves 8%-12% in 5-shot, 5%-8% in 10 shot, and 1%-5% in 20-shot improvements, respectively, compared to the existing state-of-the-art methods.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121980984","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Review of Micro-expression Recognition based on Deep Learning 基于深度学习的微表情识别研究进展
Pub Date : 2022-07-18 DOI: 10.1109/IJCNN55064.2022.9892307
He Zhang, Hanling Zhang
Micro-expression has the characteristics of spontaneity, low intensity, and short duration, which reflects a real personal emotion. Therefore, micro-expression recognition (MER) has been applied widely in lie detection, depression analysis, human-computer interaction systems, and commercial negotiation. Micro-expressions usually occur when people attempt to cover up their true feelings, especially in high-stake environments. In the early stage, the study of micro-expressions was mainly from a psychological point of view and required a very specialized skill. MER based on deep learning is a hot research direction recently, which generally includes several stages, such as image preprocessing, feature extraction, and emotion classification. In this paper, we first introduce the problems and challenges MER encountered. Then we present the commonly used micro-expression datasets and methods of image preprocessing. Next, we describe the MER methods based on deep learning in recent years and classify them according to the network structure. Afterward, we present the evaluation metrics and protocol and compare different algorithms on the composite dataset. Finally, we conclude and provide a prospect of the future work of MER.
微表情具有自发性、强度低、持续时间短的特点,反映了真实的个人情绪。因此,微表情识别在测谎、抑郁分析、人机交互系统、商业谈判等领域得到了广泛的应用。当人们试图掩盖自己的真实感受时,尤其是在高风险的环境中,通常会出现微表情。在早期,对微表情的研究主要是从心理学的角度出发,需要非常专业的技能。基于深度学习的人工神经网络是近年来的一个热点研究方向,一般包括图像预处理、特征提取、情感分类等几个阶段。在本文中,我们首先介绍了MER遇到的问题和挑战。然后介绍了常用的微表情数据集和图像预处理方法。其次,我们描述了近年来基于深度学习的MER方法,并根据网络结构对它们进行了分类。然后,我们提出了评估指标和协议,并比较了不同的算法在复合数据集上。最后,对未来的研究工作进行了总结和展望。
{"title":"A Review of Micro-expression Recognition based on Deep Learning","authors":"He Zhang, Hanling Zhang","doi":"10.1109/IJCNN55064.2022.9892307","DOIUrl":"https://doi.org/10.1109/IJCNN55064.2022.9892307","url":null,"abstract":"Micro-expression has the characteristics of spontaneity, low intensity, and short duration, which reflects a real personal emotion. Therefore, micro-expression recognition (MER) has been applied widely in lie detection, depression analysis, human-computer interaction systems, and commercial negotiation. Micro-expressions usually occur when people attempt to cover up their true feelings, especially in high-stake environments. In the early stage, the study of micro-expressions was mainly from a psychological point of view and required a very specialized skill. MER based on deep learning is a hot research direction recently, which generally includes several stages, such as image preprocessing, feature extraction, and emotion classification. In this paper, we first introduce the problems and challenges MER encountered. Then we present the commonly used micro-expression datasets and methods of image preprocessing. Next, we describe the MER methods based on deep learning in recent years and classify them according to the network structure. Afterward, we present the evaluation metrics and protocol and compare different algorithms on the composite dataset. Finally, we conclude and provide a prospect of the future work of MER.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122246936","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Foreground-attention in neural decoding: Guiding Loop-Enc-Dec to reconstruct visual stimulus images from fMRI 神经解码中的前景注意:引导环路c- dec重构fMRI视觉刺激图像
Pub Date : 2022-07-18 DOI: 10.1109/IJCNN55064.2022.9892276
Kai Chen, Yongqiang Ma, Mingyang Sheng, N. Zheng
The reconstruction of visual stimulus images from functional Magnetic Resonance Imaging (fMRI) has received extensive attention in recent years, which provides a possibility to interpret the human brain. Due to the high-dimensional and high-noise characteristics of fMRI data, how to extract stable, reliable and useful information from fMRI data for image reconstruction has become a challenging problem. Inspired by the mechanism of human visual attention, in this paper, we propose a novel method of reconstructing visual stimulus images, which first decodes human visual salient region from fMRI, we define human visual salient region as foreground attention (F-attention), and then reconstructs the visual images guided by F-attention. Because the human brain is strongly wound into sulci and gyri, some spatially adjacent voxels are not connected in practice. Therefore, it is necessary to consider the global information when decoding fMRI, so we introduce the self-attention module for capturing global information into the process of decoding F-attention. In addition, in order to obtain more loss constraints in the training process of encoder-decoder, we also propose a new training strategy called Loop-Enc-Dec. The experimental results show that the F-attention decoder decodes the visual attention from fMRI successfully, and the Loop-Enc-Dec guided by F-attention can also well reconstruct the visual stimulus images.
近年来,功能磁共振成像(fMRI)对视觉刺激图像的重建受到了广泛关注,为解释人类大脑提供了一种可能。由于功能磁共振成像数据具有高维、高噪声的特点,如何从功能磁共振成像数据中提取稳定、可靠、有用的信息进行图像重建成为一个具有挑战性的问题。受人类视觉注意机制的启发,本文提出了一种重建视觉刺激图像的新方法,该方法首先从功能磁共振成像(fMRI)中解码人类视觉显著区域,将人类视觉显著区域定义为前景注意(f -注意),然后在f -注意的引导下重建视觉图像。由于人脑被强烈地缠绕成脑沟和脑回,一些空间上相邻的体素在实践中是不相连的。因此,在fMRI解码时需要考虑全局信息,因此我们在f-注意解码过程中引入自注意模块,用于捕获全局信息。此外,为了在编码器-解码器的训练过程中获得更多的损失约束,我们还提出了一种新的训练策略loop - c- dec。实验结果表明,f -注意解码器成功解码了fMRI的视觉注意,f -注意引导下的loop - c- dec也能很好地重建视觉刺激图像。
{"title":"Foreground-attention in neural decoding: Guiding Loop-Enc-Dec to reconstruct visual stimulus images from fMRI","authors":"Kai Chen, Yongqiang Ma, Mingyang Sheng, N. Zheng","doi":"10.1109/IJCNN55064.2022.9892276","DOIUrl":"https://doi.org/10.1109/IJCNN55064.2022.9892276","url":null,"abstract":"The reconstruction of visual stimulus images from functional Magnetic Resonance Imaging (fMRI) has received extensive attention in recent years, which provides a possibility to interpret the human brain. Due to the high-dimensional and high-noise characteristics of fMRI data, how to extract stable, reliable and useful information from fMRI data for image reconstruction has become a challenging problem. Inspired by the mechanism of human visual attention, in this paper, we propose a novel method of reconstructing visual stimulus images, which first decodes human visual salient region from fMRI, we define human visual salient region as foreground attention (F-attention), and then reconstructs the visual images guided by F-attention. Because the human brain is strongly wound into sulci and gyri, some spatially adjacent voxels are not connected in practice. Therefore, it is necessary to consider the global information when decoding fMRI, so we introduce the self-attention module for capturing global information into the process of decoding F-attention. In addition, in order to obtain more loss constraints in the training process of encoder-decoder, we also propose a new training strategy called Loop-Enc-Dec. The experimental results show that the F-attention decoder decodes the visual attention from fMRI successfully, and the Loop-Enc-Dec guided by F-attention can also well reconstruct the visual stimulus images.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":"70 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115796356","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Measuring Drift Severity by Tree Structure Classifiers 用树结构分类器测量漂移严重程度
Pub Date : 2022-07-18 DOI: 10.1109/IJCNN55064.2022.9892439
Di Zhao, Yun Sing Koh, Philippe Fournier-Viger
Streaming data has become more common as our ability to collect data in real-time increases. A primary concern in dealing with data streams is concept drift, which describes changes in the underlying distribution of streaming data. Measuring drift severity is crucial for model adaptation. Drift severity can be a proxy in choosing concept drift adaptation strategies. Current methods measure drift severity by monitoring the changes in the learner performance or measuring the difference between data distributions. However, these methods cannot measure the drift severity if the ground truth labels are unavailable. Specifically, performance-based methods cannot measure marginal drift, and distribution-based methods cannot measure conditional drift. We propose a novel framework named Tree-based Drift Measurement (TDM) that measures both marginal and conditional drift without revisiting historical data. TDM measures the difference between tree classifiers by transforming them into sets of binary vectors. An experiment shows that TDM achieves similar performance to the state-of-the-art methods and provides the best trade-off between runtime and memory usage. A case study shows that the online learner performance can be improved by adapting different drift adaptation strategies based on the drift severity.
随着我们实时收集数据的能力的提高,流数据变得越来越普遍。处理数据流的一个主要关注点是概念漂移,它描述了流数据底层分布的变化。测量漂移严重程度对模型适应至关重要。漂移严重程度可以作为选择概念漂移适应策略的一个指标。目前的方法通过监测学习者表现的变化或测量数据分布之间的差异来测量漂移的严重程度。然而,如果地面真值标签不可用,这些方法无法测量漂移的严重程度。具体来说,基于性能的方法不能测量边际漂移,而基于分布的方法不能测量条件漂移。我们提出了一种新的框架,称为基于树的漂移测量(TDM),测量边际和条件漂移而无需重述历史数据。TDM通过将树分类器转换成二进制向量集来度量树分类器之间的差异。实验表明,TDM实现了与最先进的方法相似的性能,并提供了运行时和内存使用之间的最佳折衷。实例研究表明,根据漂移的严重程度,采用不同的漂移适应策略可以提高在线学习者的学习成绩。
{"title":"Measuring Drift Severity by Tree Structure Classifiers","authors":"Di Zhao, Yun Sing Koh, Philippe Fournier-Viger","doi":"10.1109/IJCNN55064.2022.9892439","DOIUrl":"https://doi.org/10.1109/IJCNN55064.2022.9892439","url":null,"abstract":"Streaming data has become more common as our ability to collect data in real-time increases. A primary concern in dealing with data streams is concept drift, which describes changes in the underlying distribution of streaming data. Measuring drift severity is crucial for model adaptation. Drift severity can be a proxy in choosing concept drift adaptation strategies. Current methods measure drift severity by monitoring the changes in the learner performance or measuring the difference between data distributions. However, these methods cannot measure the drift severity if the ground truth labels are unavailable. Specifically, performance-based methods cannot measure marginal drift, and distribution-based methods cannot measure conditional drift. We propose a novel framework named Tree-based Drift Measurement (TDM) that measures both marginal and conditional drift without revisiting historical data. TDM measures the difference between tree classifiers by transforming them into sets of binary vectors. An experiment shows that TDM achieves similar performance to the state-of-the-art methods and provides the best trade-off between runtime and memory usage. A case study shows that the online learner performance can be improved by adapting different drift adaptation strategies based on the drift severity.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131348774","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Feedforward neural networks in forecasting the spatial distribution of the time-dependent multidimensional functions 前馈神经网络在预测时变多维函数空间分布中的应用
Pub Date : 2022-07-18 DOI: 10.1109/IJCNN55064.2022.9892001
A. Wawrzynczak, M. Berendt-Marchel
The neural networks are powerful as nonlinear signal processors. This paper deals with the problem of employing the feedforward neural networks (FFNNs) to simulate the time-dependent distribution of the airborne toxin in the urbanized area. The spatial distribution of the contaminant is the multidimensional function dependent on the weather conditions (wind direction and speed), coordinates of the contamination sources, the release rate, and its duration. In this paper, we try to answer what topology should be the multilayered FFNN to forecast the contaminant strength correctly at the given point of the urbanized area at a given time. The comparison between the FFNNs is made based on the standard performance measures like correlation R and mean square error (MSE). Additionally, the new measure estimating the quality of the neural networks forecasts in subsequent time intervals after the release is proposed. In combination with R and MSE, the proposed measure allows identifying the well-trained network unambiguously. Such a neural network may enable creating an emergency system localizing the contaminant source in an urban area in real-time. However, in such a system time of answer depends directly on the multiple times run dispersion model computational time. This time is expected in minutes for custom dispersion models in urban areas and can be shortened to seconds in the case of artificial neural networks.
神经网络是一种强大的非线性信号处理器。本文研究了利用前馈神经网络(FFNNs)模拟城市化地区空气中毒素随时间分布的问题。污染物的空间分布是依赖于天气条件(风向和风速)、污染源坐标、释放速率及其持续时间的多维函数。在本文中,我们试图回答多层FFNN应该是什么拓扑结构,以正确预测在给定时间的城市化区域的给定点的污染物强度。基于相关R和均方误差(MSE)等标准性能指标对ffnn进行比较。此外,还提出了一种新的方法来估计神经网络在发布后的后续时间间隔内的预测质量。结合R和MSE,所提出的方法可以明确地识别训练良好的网络。这种神经网络可以创建一个实时定位城市地区污染源的应急系统。然而,在这样的系统中,答案的时间直接取决于多次运行色散模型的计算时间。对于城市地区的自定义分散模型,这一时间预计为几分钟,而对于人工神经网络,这一时间可以缩短到几秒钟。
{"title":"Feedforward neural networks in forecasting the spatial distribution of the time-dependent multidimensional functions","authors":"A. Wawrzynczak, M. Berendt-Marchel","doi":"10.1109/IJCNN55064.2022.9892001","DOIUrl":"https://doi.org/10.1109/IJCNN55064.2022.9892001","url":null,"abstract":"The neural networks are powerful as nonlinear signal processors. This paper deals with the problem of employing the feedforward neural networks (FFNNs) to simulate the time-dependent distribution of the airborne toxin in the urbanized area. The spatial distribution of the contaminant is the multidimensional function dependent on the weather conditions (wind direction and speed), coordinates of the contamination sources, the release rate, and its duration. In this paper, we try to answer what topology should be the multilayered FFNN to forecast the contaminant strength correctly at the given point of the urbanized area at a given time. The comparison between the FFNNs is made based on the standard performance measures like correlation R and mean square error (MSE). Additionally, the new measure estimating the quality of the neural networks forecasts in subsequent time intervals after the release is proposed. In combination with R and MSE, the proposed measure allows identifying the well-trained network unambiguously. Such a neural network may enable creating an emergency system localizing the contaminant source in an urban area in real-time. However, in such a system time of answer depends directly on the multiple times run dispersion model computational time. This time is expected in minutes for custom dispersion models in urban areas and can be shortened to seconds in the case of artificial neural networks.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131456451","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2022 International Joint Conference on Neural Networks (IJCNN)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1