Detection of depression in speech

Zhenyu Liu, B. Hu, Lihua Yan, Tian-Zhong Wang, Fei Liu, Xiaoyu Li, Huanyu Kang
{"title":"Detection of depression in speech","authors":"Zhenyu Liu, B. Hu, Lihua Yan, Tian-Zhong Wang, Fei Liu, Xiaoyu Li, Huanyu Kang","doi":"10.1109/ACII.2015.7344652","DOIUrl":null,"url":null,"abstract":"Depression is a common mental disorder and one of the main causes of disability worldwide. Lacking objective depressive disorder assessment methods is the key reason that many depressive patients can't be treated properly. Developments in affective sensing technology with a focus on acoustic features will potentially bring a change due to depressed patient's slow, hesitating, monotonous voice as remarkable characteristics. Our motivation is to find out a speech feature set to detect, evaluate and even predict depression. For these goals, we investigate a large sample of 300 subjects (100 depressed patients, 100 healthy controls and 100 high-risk people) through comparative analysis and follow-up study. For examining the correlation between depression and speech, we extract features as many as possible according to previous research to create a large voice feature set. Then we employ some feature selection methods to eliminate irrelevant, redundant and noisy features to form a compact subset. To measure effectiveness of this new subset, we test it on our dataset with 300 subjects using several common classifiers and 10-fold cross-validation. Since we are collecting data currently, we have no result to report yet.","PeriodicalId":6863,"journal":{"name":"2015 International Conference on Affective Computing and Intelligent Interaction (ACII)","volume":"32 1","pages":"743-747"},"PeriodicalIF":0.0000,"publicationDate":"2015-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"34","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2015 International Conference on Affective Computing and Intelligent Interaction (ACII)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ACII.2015.7344652","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 34

Abstract

Depression is a common mental disorder and one of the main causes of disability worldwide. Lacking objective depressive disorder assessment methods is the key reason that many depressive patients can't be treated properly. Developments in affective sensing technology with a focus on acoustic features will potentially bring a change due to depressed patient's slow, hesitating, monotonous voice as remarkable characteristics. Our motivation is to find out a speech feature set to detect, evaluate and even predict depression. For these goals, we investigate a large sample of 300 subjects (100 depressed patients, 100 healthy controls and 100 high-risk people) through comparative analysis and follow-up study. For examining the correlation between depression and speech, we extract features as many as possible according to previous research to create a large voice feature set. Then we employ some feature selection methods to eliminate irrelevant, redundant and noisy features to form a compact subset. To measure effectiveness of this new subset, we test it on our dataset with 300 subjects using several common classifiers and 10-fold cross-validation. Since we are collecting data currently, we have no result to report yet.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
言语抑制性的检测
抑郁症是一种常见的精神障碍,也是全世界致残的主要原因之一。缺乏客观的抑郁障碍评估方法是许多抑郁症患者无法得到正确治疗的关键原因。以声学特征为重点的情感传感技术的发展将有可能改变抑郁症患者缓慢、犹豫、单调的声音作为显著特征。我们的动机是找到一个语音特征集来检测、评估甚至预测抑郁症。为此,我们对300名受试者(100名抑郁症患者、100名健康对照者和100名高危人群)进行了比较分析和随访研究。为了检验抑郁和语音之间的相关性,我们根据之前的研究尽可能多地提取特征,以创建一个大的语音特征集。然后采用一些特征选择方法去除不相关的、冗余的和有噪声的特征,形成一个紧凑的子集。为了衡量这个新子集的有效性,我们使用几个常见分类器和10倍交叉验证在我们的数据集上测试了300个主题。由于我们目前正在收集数据,所以还没有结果可以报告。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Avatar and participant gender differences in the perception of uncanniness of virtual humans Neural conditional ordinal random fields for agreement level estimation Fundamental frequency modeling using wavelets for emotional voice conversion Bimodal feature-based fusion for real-time emotion recognition in a mobile context Harmony search for feature selection in speech emotion recognition
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1