Quantifying Cognitive Load from Voice using Transformer-Based Models and a Cross-Dataset Evaluation

Pascal Hecker, A. Kappattanavar, Maximilian Schmitt, S. Moontaha, Johannes Wagner, F. Eyben, Björn Schuller, B. Arnrich
{"title":"Quantifying Cognitive Load from Voice using Transformer-Based Models and a Cross-Dataset Evaluation","authors":"Pascal Hecker, A. Kappattanavar, Maximilian Schmitt, S. Moontaha, Johannes Wagner, F. Eyben, Björn Schuller, B. Arnrich","doi":"10.1109/ICMLA55696.2022.00055","DOIUrl":null,"url":null,"abstract":"Cognitive load is frequently induced in laboratory setups to measure responses to stress, and its impact on voice has been studied in the field of computational paralinguistics. One dataset on this topic was provided in the Computational Paralinguistics Challenge (ComParE) 2014, and therefore offers great comparability. Recently, transformer-based deep learning architectures established a new state-of-the-art and are finding their way gradually into the audio domain. In this context, we investigate the performance of popular transformer architectures in the audio domain on the ComParE 2014 dataset, and the impact of different pre-training and fine-tuning setups on these models. Further, we recorded a small custom dataset, designed to be comparable with the ComParE 2014 one, to assess cross-corpus model generalisability. We find that the transformer models outperform the challenge baseline, the challenge winner, and more recent deep learning approaches. Models based on the ‘large’ architecture perform well on the task at hand, while models based on the ‘base’ architecture perform at chance level. Fine-tuning on related domains (such as ASR or emotion), before fine-tuning on the targets, yields no higher performance compared to models pre-trained only in a self-supervised manner. The generalisability of the models between datasets is more intricate than expected, as seen in an unexpected low performance on the small custom dataset, and we discuss potential ‘hidden’ underlying discrepancies between the datasets. In summary, transformer-based architectures outperform previous attempts to quantify cognitive load from voice. This is promising, in particular for healthcare-related problems in computational paralinguistics applications, since datasets are sparse in that realm.","PeriodicalId":128160,"journal":{"name":"2022 21st IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 21st IEEE International Conference on Machine Learning and Applications (ICMLA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICMLA55696.2022.00055","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Cognitive load is frequently induced in laboratory setups to measure responses to stress, and its impact on voice has been studied in the field of computational paralinguistics. One dataset on this topic was provided in the Computational Paralinguistics Challenge (ComParE) 2014, and therefore offers great comparability. Recently, transformer-based deep learning architectures established a new state-of-the-art and are finding their way gradually into the audio domain. In this context, we investigate the performance of popular transformer architectures in the audio domain on the ComParE 2014 dataset, and the impact of different pre-training and fine-tuning setups on these models. Further, we recorded a small custom dataset, designed to be comparable with the ComParE 2014 one, to assess cross-corpus model generalisability. We find that the transformer models outperform the challenge baseline, the challenge winner, and more recent deep learning approaches. Models based on the ‘large’ architecture perform well on the task at hand, while models based on the ‘base’ architecture perform at chance level. Fine-tuning on related domains (such as ASR or emotion), before fine-tuning on the targets, yields no higher performance compared to models pre-trained only in a self-supervised manner. The generalisability of the models between datasets is more intricate than expected, as seen in an unexpected low performance on the small custom dataset, and we discuss potential ‘hidden’ underlying discrepancies between the datasets. In summary, transformer-based architectures outperform previous attempts to quantify cognitive load from voice. This is promising, in particular for healthcare-related problems in computational paralinguistics applications, since datasets are sparse in that realm.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于转换模型和跨数据集评估的语音认知负荷量化
在实验室环境中,认知负荷经常被用来测量对压力的反应,它对语音的影响已经在计算副语言学领域进行了研究。关于这个主题的一个数据集是在2014年计算副语言学挑战(ComParE)中提供的,因此具有很强的可比性。最近,基于转换器的深度学习架构建立了一种新的技术,并逐渐进入音频领域。在此背景下,我们在ComParE 2014数据集上研究了音频领域中流行的变压器架构的性能,以及不同的预训练和微调设置对这些模型的影响。此外,我们记录了一个小型自定义数据集,旨在与ComParE 2014的数据集进行比较,以评估跨语料库模型的通用性。我们发现,转换模型优于挑战基线、挑战赢家和最近的深度学习方法。基于“大型”架构的模型在手头的任务上表现良好,而基于“基础”架构的模型在偶然级别上表现良好。在对目标进行微调之前,对相关领域(如ASR或情感)进行微调,与仅以自我监督的方式预训练的模型相比,不会产生更高的性能。数据集之间模型的通用性比预期的更复杂,正如在小型自定义数据集上意想不到的低性能所看到的那样,我们讨论了数据集之间潜在的“隐藏”潜在差异。总之,基于转换器的体系结构优于以前量化语音认知负荷的尝试。这是很有希望的,特别是对于计算副语言学应用中的医疗保健相关问题,因为该领域的数据集是稀疏的。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Approximate Orthogonal Spectral Autoencoders for Community Analysis in Social Networks DeepReject and DeepRoad: Road Condition Recognition and Classification Under Adversarial Conditions Improving Aquaculture Systems using AI: Employing predictive models for Biomass Estimation on Sonar Images ICDARTS: Improving the Stability of Cyclic DARTS Symbolic Semantic Memory in Transformer Language Models
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1