不同年龄组基于面部表情的情绪识别:采用对比学习方法的多尺度视觉转换器

IF 0.9 4区 数学 Q4 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Journal of Combinatorial Optimization Pub Date : 2024-12-16 DOI:10.1007/s10878-024-01241-8
G. Balachandran, S. Ranjith, T. R. Chenthil, G. C. Jagan
{"title":"不同年龄组基于面部表情的情绪识别:采用对比学习方法的多尺度视觉转换器","authors":"G. Balachandran, S. Ranjith, T. R. Chenthil, G. C. Jagan","doi":"10.1007/s10878-024-01241-8","DOIUrl":null,"url":null,"abstract":"<p>Facial expression-based Emotion Recognition (FER) is crucial in human–computer interaction and affective computing, particularly when addressing diverse age groups. This paper introduces the Multi-Scale Vision Transformer with Contrastive Learning (MViT-CnG), an age-adaptive FER approach designed to enhance the accuracy and interpretability of emotion recognition models across different classes. The MViT-CnG model leverages vision transformers and contrastive learning to capture intricate facial features, ensuring robust performance despite diverse and dynamic facial features. By utilizing contrastive learning, the model's interpretability is significantly enhanced, which is vital for building trust in automated systems and facilitating human–machine collaboration. Additionally, this approach enriches the model's capacity to discern shared and distinct features within facial expressions, improving its ability to generalize across different age groups. Evaluations using the FER-2013 and CK + datasets highlight the model's broad generalization capabilities, with FER-2013 covering a wide range of emotions across diverse age groups and CK + focusing on posed expressions in controlled environments. The MViT-CnG model adapts effectively to both datasets, showcasing its versatility and reliability across distinct data characteristics. Performance results demonstrated that the MViT-CnG model achieved superior accuracy across all emotion recognition labels on the FER-2013 dataset with a 99.6% accuracy rate, and 99.5% on the CK + dataset, indicating significant improvements in recognizing subtle facial expressions. Comprehensive evaluations revealed that the model's precision, recall, and F1-score are consistently higher than those of existing models, confirming its robustness and reliability in facial emotion recognition tasks.</p>","PeriodicalId":50231,"journal":{"name":"Journal of Combinatorial Optimization","volume":"21 1","pages":""},"PeriodicalIF":0.9000,"publicationDate":"2024-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Facial expression-based emotion recognition across diverse age groups: a multi-scale vision transformer with contrastive learning approach\",\"authors\":\"G. Balachandran, S. Ranjith, T. R. Chenthil, G. C. Jagan\",\"doi\":\"10.1007/s10878-024-01241-8\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>Facial expression-based Emotion Recognition (FER) is crucial in human–computer interaction and affective computing, particularly when addressing diverse age groups. This paper introduces the Multi-Scale Vision Transformer with Contrastive Learning (MViT-CnG), an age-adaptive FER approach designed to enhance the accuracy and interpretability of emotion recognition models across different classes. The MViT-CnG model leverages vision transformers and contrastive learning to capture intricate facial features, ensuring robust performance despite diverse and dynamic facial features. By utilizing contrastive learning, the model's interpretability is significantly enhanced, which is vital for building trust in automated systems and facilitating human–machine collaboration. Additionally, this approach enriches the model's capacity to discern shared and distinct features within facial expressions, improving its ability to generalize across different age groups. Evaluations using the FER-2013 and CK + datasets highlight the model's broad generalization capabilities, with FER-2013 covering a wide range of emotions across diverse age groups and CK + focusing on posed expressions in controlled environments. The MViT-CnG model adapts effectively to both datasets, showcasing its versatility and reliability across distinct data characteristics. Performance results demonstrated that the MViT-CnG model achieved superior accuracy across all emotion recognition labels on the FER-2013 dataset with a 99.6% accuracy rate, and 99.5% on the CK + dataset, indicating significant improvements in recognizing subtle facial expressions. Comprehensive evaluations revealed that the model's precision, recall, and F1-score are consistently higher than those of existing models, confirming its robustness and reliability in facial emotion recognition tasks.</p>\",\"PeriodicalId\":50231,\"journal\":{\"name\":\"Journal of Combinatorial Optimization\",\"volume\":\"21 1\",\"pages\":\"\"},\"PeriodicalIF\":0.9000,\"publicationDate\":\"2024-12-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Combinatorial Optimization\",\"FirstCategoryId\":\"100\",\"ListUrlMain\":\"https://doi.org/10.1007/s10878-024-01241-8\",\"RegionNum\":4,\"RegionCategory\":\"数学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q4\",\"JCRName\":\"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Combinatorial Optimization","FirstCategoryId":"100","ListUrlMain":"https://doi.org/10.1007/s10878-024-01241-8","RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
引用次数: 0

摘要

基于面部表情的情感识别(FER)在人机交互和情感计算中至关重要,特别是在处理不同年龄组时。本文介绍了基于对比学习的多尺度视觉变换(Multi-Scale Vision Transformer with Contrastive Learning, mviti - cng),这是一种年龄自适应的FER方法,旨在提高不同类别情绪识别模型的准确性和可解释性。MViT-CnG模型利用视觉变压器和对比学习来捕获复杂的面部特征,确保在多样化和动态的面部特征下仍然具有强大的性能。通过使用对比学习,模型的可解释性显著增强,这对于在自动化系统中建立信任和促进人机协作至关重要。此外,这种方法丰富了模型在面部表情中识别共同特征和不同特征的能力,提高了它在不同年龄段进行概括的能力。使用FER-2013和CK +数据集的评估突出了模型的广泛泛化能力,其中FER-2013涵盖了不同年龄组的广泛情绪,而CK +侧重于受控环境中的姿势表情。mviti - cng模型有效地适应了这两种数据集,展示了其在不同数据特征下的通用性和可靠性。性能结果表明,MViT-CnG模型在FER-2013数据集中的所有情绪识别标签上都取得了优异的准确率,准确率为99.6%,CK +数据集中的准确率为99.5%,表明在识别细微面部表情方面有显著提高。综合评价表明,该模型的准确率、召回率和f1得分均高于现有模型,证实了其在面部情绪识别任务中的鲁棒性和可靠性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Facial expression-based emotion recognition across diverse age groups: a multi-scale vision transformer with contrastive learning approach

Facial expression-based Emotion Recognition (FER) is crucial in human–computer interaction and affective computing, particularly when addressing diverse age groups. This paper introduces the Multi-Scale Vision Transformer with Contrastive Learning (MViT-CnG), an age-adaptive FER approach designed to enhance the accuracy and interpretability of emotion recognition models across different classes. The MViT-CnG model leverages vision transformers and contrastive learning to capture intricate facial features, ensuring robust performance despite diverse and dynamic facial features. By utilizing contrastive learning, the model's interpretability is significantly enhanced, which is vital for building trust in automated systems and facilitating human–machine collaboration. Additionally, this approach enriches the model's capacity to discern shared and distinct features within facial expressions, improving its ability to generalize across different age groups. Evaluations using the FER-2013 and CK + datasets highlight the model's broad generalization capabilities, with FER-2013 covering a wide range of emotions across diverse age groups and CK + focusing on posed expressions in controlled environments. The MViT-CnG model adapts effectively to both datasets, showcasing its versatility and reliability across distinct data characteristics. Performance results demonstrated that the MViT-CnG model achieved superior accuracy across all emotion recognition labels on the FER-2013 dataset with a 99.6% accuracy rate, and 99.5% on the CK + dataset, indicating significant improvements in recognizing subtle facial expressions. Comprehensive evaluations revealed that the model's precision, recall, and F1-score are consistently higher than those of existing models, confirming its robustness and reliability in facial emotion recognition tasks.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Journal of Combinatorial Optimization
Journal of Combinatorial Optimization 数学-计算机:跨学科应用
CiteScore
2.00
自引率
10.00%
发文量
83
审稿时长
6 months
期刊介绍: The objective of Journal of Combinatorial Optimization is to advance and promote the theory and applications of combinatorial optimization, which is an area of research at the intersection of applied mathematics, computer science, and operations research and which overlaps with many other areas such as computation complexity, computational biology, VLSI design, communication networks, and management science. It includes complexity analysis and algorithm design for combinatorial optimization problems, numerical experiments and problem discovery with applications in science and engineering. The Journal of Combinatorial Optimization publishes refereed papers dealing with all theoretical, computational and applied aspects of combinatorial optimization. It also publishes reviews of appropriate books and special issues of journals.
期刊最新文献
Social media actors: perception and optimization of influence across different types $$(K_{1}\vee {P_{t})}$$ -saturated graphs with minimum number of edges Advertising meets assortment planning: joint advertising and assortment optimization under multinomial logit model Recognizing integrality of weighted rectangles partitions An improved PTAS for covering targets with mobile sensors
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1