Investigation of unsupervised adaptation of DNN acoustic models with filter bank input

Takuya Yoshioka, A. Ragni, M. Gales
{"title":"Investigation of unsupervised adaptation of DNN acoustic models with filter bank input","authors":"Takuya Yoshioka, A. Ragni, M. Gales","doi":"10.1109/ICASSP.2014.6854825","DOIUrl":null,"url":null,"abstract":"Adaptation to speaker variations is an essential component of speech recognition systems. One common approach to adapting deep neural network (DNN) acoustic models is to perform global constrained maximum likelihood linear regression (CMLLR) at some point of the systems. Using CMLLR (or more generally, generative approaches) is advantageous especially in unsupervised adaptation scenarios with high baseline error rates. On the other hand, as the DNNs are less sensitive to the increase in the input dimensionality than GMMs, it is becoming more popular to use rich speech representations, such as log mel-filter bank channel outputs, instead of conventional low-dimensional feature vectors, such as MFCCs and PLP coefficients. This work discusses and compares three different configurations of DNN acoustic models that allow CMLLR-based speaker adaptive training (SAT) to be performed in systems with filter bank inputs. Results of unsupervised adaptation experiments conducted on three different data sets are presented, demonstrating that, by choosing an appropriate configuration, SAT with CMLLR can improve the performance of a well-trained filter bank-based speaker independent DNN system by 10.6% relative in a challenging task with a baseline error rate above 40%. It is also shown that the filter bank features are advantageous than the conventional features even when they are used with SAT models. Some other insights are also presented, including the effects of block diagonal transforms and system combination.","PeriodicalId":6545,"journal":{"name":"2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"13 1","pages":"6344-6348"},"PeriodicalIF":0.0000,"publicationDate":"2014-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"42","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICASSP.2014.6854825","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 42

Abstract

Adaptation to speaker variations is an essential component of speech recognition systems. One common approach to adapting deep neural network (DNN) acoustic models is to perform global constrained maximum likelihood linear regression (CMLLR) at some point of the systems. Using CMLLR (or more generally, generative approaches) is advantageous especially in unsupervised adaptation scenarios with high baseline error rates. On the other hand, as the DNNs are less sensitive to the increase in the input dimensionality than GMMs, it is becoming more popular to use rich speech representations, such as log mel-filter bank channel outputs, instead of conventional low-dimensional feature vectors, such as MFCCs and PLP coefficients. This work discusses and compares three different configurations of DNN acoustic models that allow CMLLR-based speaker adaptive training (SAT) to be performed in systems with filter bank inputs. Results of unsupervised adaptation experiments conducted on three different data sets are presented, demonstrating that, by choosing an appropriate configuration, SAT with CMLLR can improve the performance of a well-trained filter bank-based speaker independent DNN system by 10.6% relative in a challenging task with a baseline error rate above 40%. It is also shown that the filter bank features are advantageous than the conventional features even when they are used with SAT models. Some other insights are also presented, including the effects of block diagonal transforms and system combination.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
带滤波器组输入的DNN声学模型的无监督自适应研究
对说话人变化的适应是语音识别系统的重要组成部分。自适应深度神经网络(DNN)声学模型的一种常用方法是在系统的某个点执行全局约束最大似然线性回归(CMLLR)。使用cmlr(或者更一般地说,生成方法)是有利的,特别是在具有高基线错误率的无监督适应场景中。另一方面,由于dnn对输入维数的增加不像gmm那么敏感,因此使用丰富的语音表示(如对数梅尔滤波器组通道输出)而不是传统的低维特征向量(如mfccc和PLP系数)变得越来越流行。本研究讨论并比较了DNN声学模型的三种不同配置,这些模型允许基于cmlr的扬声器自适应训练(SAT)在具有滤波器组输入的系统中执行。在三个不同的数据集上进行的无监督自适应实验结果表明,通过选择适当的配置,SAT与cmlr可以将训练良好的基于滤波器组的说话人独立DNN系统的性能提高10.6%,相对于基线错误率在40%以上的挑战性任务。结果还表明,即使与SAT模型一起使用,滤波器组特征也比传统特征更有优势。本文还提出了一些其他的见解,包括块对角变换和系统组合的影响。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Multichannel detection of an unknown rank-one signal with uncalibrated receivers Design and implementation of a low power spike detection processor for 128-channel spike sorting microsystem On the convergence of average consensus with generalized metropolis-hasting weights A network of HF surface wave radars for maritime surveillance: Preliminary results in the German Bight Mobile real-time arousal detection
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1