A lightweight and privacy preserved federated learning ecosystem for analyzing verbal communication emotions in identical and non-identical databases

Q4 Engineering Measurement Sensors Pub Date : 2024-06-26 DOI:10.1016/j.measen.2024.101268
Muskan Chawla , Surya Narayan Panda , Vikas Khullar , Sushil Kumar , Shyama Barna Bhattacharjee
{"title":"A lightweight and privacy preserved federated learning ecosystem for analyzing verbal communication emotions in identical and non-identical databases","authors":"Muskan Chawla ,&nbsp;Surya Narayan Panda ,&nbsp;Vikas Khullar ,&nbsp;Sushil Kumar ,&nbsp;Shyama Barna Bhattacharjee","doi":"10.1016/j.measen.2024.101268","DOIUrl":null,"url":null,"abstract":"<div><p>The lack of vocal emotional expression is a major deficit in social communication disorders. The current scenario of artificial intelligence focuses on collaborative training of deep learning models without losing data privacy. The primary objective of this paper is to propose a federated learning-based classification model to identify and analyze the emotional capabilities of individuals with vocal emotion deficits. The methodology has developed a collaborative and privacy-preserved approach using federated learning for training the deep learning models. The proposed methodology utilizes Mel-frequency Cepstral Coefficients (MFCC) to preprocess audio recordings. The four datasets (RAVDESS, CREMA, TESS, SAVEE) including emotion-based classified audio recordings were collected from open sources. The collected audio recordings are 3 s each and the total data set has 668376 audio files with happy - 175119 files, sad – 172611 files, angry – 176346 files, and normal - 144300 files. Further, the input audio was pre-processed to generate MFCC features. The study began with extracting features from multiple pre-trained DL models as its base model. Then, the performance of the federated learning (FL) model was tested on independent and identically distributed (IID) and non-IID data. Further, this paper presents a federated deep learning-based multimodal system for verbal communication emotions classification that uses audio datasets to meet data privacy requirements by DL on the FL ecosystem. As per the findings, the federated learning trained model provides nearly similar parametric results in comparison to base model training. For IID data, the model had 99.71 % validation accuracy, precision (99.73 %), recall (99.69 %), and validation loss (0.01). The FL architecture with non-IID data outperformed these measures with validation accuracy (99.97 %), precision (99.97 %), recall (99.97 %), and least loss (0). Hence the acquired results support the utilization of federated learning ecosystem-based trained models with identically and non-identically distributed audio features from emotion identification without losing parametric results. In conclusion, the proposed techniques could be applied to identify verbal emotional deficits in individuals and could support developing emerging technological interventions for their well-being.</p></div>","PeriodicalId":34311,"journal":{"name":"Measurement Sensors","volume":"34 ","pages":"Article 101268"},"PeriodicalIF":0.0000,"publicationDate":"2024-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2665917424002447/pdfft?md5=79b1545a9b1bd69e40c9a926a8243d17&pid=1-s2.0-S2665917424002447-main.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Measurement Sensors","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2665917424002447","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"Engineering","Score":null,"Total":0}
引用次数: 0

Abstract

The lack of vocal emotional expression is a major deficit in social communication disorders. The current scenario of artificial intelligence focuses on collaborative training of deep learning models without losing data privacy. The primary objective of this paper is to propose a federated learning-based classification model to identify and analyze the emotional capabilities of individuals with vocal emotion deficits. The methodology has developed a collaborative and privacy-preserved approach using federated learning for training the deep learning models. The proposed methodology utilizes Mel-frequency Cepstral Coefficients (MFCC) to preprocess audio recordings. The four datasets (RAVDESS, CREMA, TESS, SAVEE) including emotion-based classified audio recordings were collected from open sources. The collected audio recordings are 3 s each and the total data set has 668376 audio files with happy - 175119 files, sad – 172611 files, angry – 176346 files, and normal - 144300 files. Further, the input audio was pre-processed to generate MFCC features. The study began with extracting features from multiple pre-trained DL models as its base model. Then, the performance of the federated learning (FL) model was tested on independent and identically distributed (IID) and non-IID data. Further, this paper presents a federated deep learning-based multimodal system for verbal communication emotions classification that uses audio datasets to meet data privacy requirements by DL on the FL ecosystem. As per the findings, the federated learning trained model provides nearly similar parametric results in comparison to base model training. For IID data, the model had 99.71 % validation accuracy, precision (99.73 %), recall (99.69 %), and validation loss (0.01). The FL architecture with non-IID data outperformed these measures with validation accuracy (99.97 %), precision (99.97 %), recall (99.97 %), and least loss (0). Hence the acquired results support the utilization of federated learning ecosystem-based trained models with identically and non-identically distributed audio features from emotion identification without losing parametric results. In conclusion, the proposed techniques could be applied to identify verbal emotional deficits in individuals and could support developing emerging technological interventions for their well-being.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
用于分析相同和非相同数据库中语言交流情感的轻量级、保护隐私的联合学习生态系统
缺乏声音情感表达是社交沟通障碍的一大缺陷。当前人工智能的应用场景主要是在不丢失数据隐私的前提下对深度学习模型进行协同训练。本文的主要目的是提出一种基于联合学习的分类模型,用于识别和分析发声情感缺失者的情感能力。该方法开发了一种使用联合学习训练深度学习模型的协作和隐私保护方法。所提出的方法利用梅尔频率倒频谱系数(MFCC)对音频录音进行预处理。四个数据集(RAVDESS、CREMA、TESS、SAVEE)包括基于情感的分类音频录音,均从公开来源收集。收集到的音频记录每段 3 秒,数据集共有 668376 个音频文件,其中开心的有 175119 个文件,悲伤的有 172611 个文件,生气的有 176346 个文件,正常的有 144300 个文件。此外,还对输入音频进行了预处理,以生成 MFCC 特征。研究首先从多个预训练的 DL 模型中提取特征作为基础模型。然后,在独立同分布(IID)和非独立同分布数据上测试了联合学习(FL)模型的性能。此外,本文还介绍了一种基于联合深度学习的多模态语言交流情感分类系统,该系统使用音频数据集,通过 FL 生态系统上的 DL 满足数据隐私要求。根据研究结果,与基础模型训练相比,联盟学习训练的模型提供了几乎相似的参数结果。对于 IID 数据,该模型的验证准确率为 99.71%,精确度为 99.73%,召回率为 99.69%,验证损失为 0.01%。使用非 IID 数据的 FL 架构在验证准确率(99.97 %)、精确率(99.97 %)、召回率(99.97 %)和最小损失(0)方面均优于上述指标。因此,所获得的结果支持利用基于联合学习生态系统的训练模型,在不损失参数结果的情况下,从情感识别中获得相同和非相同分布的音频特征。总之,所提出的技术可用于识别个人的言语情绪缺陷,并有助于开发新兴的技术干预措施,以促进他们的福祉。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Measurement Sensors
Measurement Sensors Engineering-Industrial and Manufacturing Engineering
CiteScore
3.10
自引率
0.00%
发文量
184
审稿时长
56 days
期刊最新文献
Vehicle maneuver recognition and correction algorithm for road quality measurement system optimization A simple and inexpensive design of radioactive source for accurate level and height gauging in petrochemical industries Pneumonia detection from X-ray images using federated learning–An unsupervised learning approach Number plate recognition smart parking management system using IoT Collaboration of clustering and classification techniques for better prediction of severity of heart stroke using deep learning
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1