{"title":"An improved recurrent neural network language model with context vector features","authors":"Jian Zhang, Dan Qu, Zhen Li","doi":"10.1109/ICSESS.2014.6933694","DOIUrl":null,"url":null,"abstract":"Recurrent neural network language models have solved the problems of data sparseness and dimensionality disaster which exist in traditional N-gram models. RNNLMs have recently demonstrated state-of-the-art performance in speech recognition, machine translation and other tasks. In this paper, we improve the model performance by providing contextual word vectors in association with RNNLMs. This method can reinforce the ability of learning long-distance information using vectors training from Skip-gram model. The experimental results show that the proposed method can improve the perplexity performance significantly on Penn Treebank data. And we further apply the models to speech recognition task on the Wall Street Journal corpora, where we achieve obvious improvements in word-error-rate.","PeriodicalId":6473,"journal":{"name":"2014 IEEE 5th International Conference on Software Engineering and Service Science","volume":"34 1","pages":"828-831"},"PeriodicalIF":0.0000,"publicationDate":"2014-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2014 IEEE 5th International Conference on Software Engineering and Service Science","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICSESS.2014.6933694","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4
Abstract
Recurrent neural network language models have solved the problems of data sparseness and dimensionality disaster which exist in traditional N-gram models. RNNLMs have recently demonstrated state-of-the-art performance in speech recognition, machine translation and other tasks. In this paper, we improve the model performance by providing contextual word vectors in association with RNNLMs. This method can reinforce the ability of learning long-distance information using vectors training from Skip-gram model. The experimental results show that the proposed method can improve the perplexity performance significantly on Penn Treebank data. And we further apply the models to speech recognition task on the Wall Street Journal corpora, where we achieve obvious improvements in word-error-rate.