首页 > 最新文献

2019 International Conference on Asian Language Processing (IALP)最新文献

英文 中文
Character Decomposition for Japanese-Chinese Character-Level Neural Machine Translation 日中字符级神经机器翻译的字符分解
Pub Date : 2019-11-01 DOI: 10.1109/IALP48816.2019.9037677
Jinyi Zhang, Tadahiro Matsumoto
After years of development, Neural Machine Translation (NMT) has produced richer translation results than ever over various language pairs, becoming a new machine translation model with great potential. For the NMT model, it can only translate words/characters contained in the training data. One problem on NMT is handling of the low-frequency words/characters in the training data. In this paper, we propose a method for removing characters whose frequencies of appearance are less than a given minimum threshold by decomposing such characters into their components and/or pseudo-characters, using the Chinese character decomposition table we made. Experiments of Japanese-to-Chinese and Chinese-to-Japanese NMT with ASPEC-JC (Asian Scientific Paper Excerpt Corpus, Japanese-Chinese) corpus show that the BLEU scores, the training time and the number of parameters are varied with the number of the given minimum thresholds of decomposed characters.
经过多年的发展,神经机器翻译(NMT)在各种语言对上的翻译结果比以往更加丰富,成为一种具有巨大潜力的新型机器翻译模型。对于NMT模型,它只能翻译训练数据中包含的单词/字符。NMT中的一个问题是训练数据中低频词/字符的处理。本文提出了一种去除出现频率小于给定最小阈值的汉字的方法,该方法使用我们制作的汉字分解表,将这些汉字分解为其组成和/或伪字符。用ASPEC-JC (Asian Scientific Paper摘录Corpus, Japanese-Chinese)语料库对日文-汉文和中文-日文NMT进行的实验表明,BLEU分数、训练时间和参数数量随给定的分解字符最小阈值的个数而变化。
{"title":"Character Decomposition for Japanese-Chinese Character-Level Neural Machine Translation","authors":"Jinyi Zhang, Tadahiro Matsumoto","doi":"10.1109/IALP48816.2019.9037677","DOIUrl":"https://doi.org/10.1109/IALP48816.2019.9037677","url":null,"abstract":"After years of development, Neural Machine Translation (NMT) has produced richer translation results than ever over various language pairs, becoming a new machine translation model with great potential. For the NMT model, it can only translate words/characters contained in the training data. One problem on NMT is handling of the low-frequency words/characters in the training data. In this paper, we propose a method for removing characters whose frequencies of appearance are less than a given minimum threshold by decomposing such characters into their components and/or pseudo-characters, using the Chinese character decomposition table we made. Experiments of Japanese-to-Chinese and Chinese-to-Japanese NMT with ASPEC-JC (Asian Scientific Paper Excerpt Corpus, Japanese-Chinese) corpus show that the BLEU scores, the training time and the number of parameters are varied with the number of the given minimum thresholds of decomposed characters.","PeriodicalId":208066,"journal":{"name":"2019 International Conference on Asian Language Processing (IALP)","volume":"98 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134187956","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
How to Answer Comparison Questions 如何回答比较问题
Pub Date : 2019-11-01 DOI: 10.1109/IALP48816.2019.9037729
Hongxuan Tang, Yu Hong, Xin Chen, Kaili Wu, Min Zhang
“Which city has the larger population, Tokyo or New York?”. To answer the question, in general, we necessarily obtain the prior knowledge about the populations of both cities, and accordingly determine the answer by numeric comparison. Using Machine Reading Comprehension (MRC) to answer such a question has become a popular research topic, which is referred to as a task of Comparison Question Answering (CQA). In this paper, we propose a novel neural CQA model which is trained to answer comparison question. The model is designed as a sophisticated neural network which performs inference in a step-by-step pipeline, including the steps of attentive entity detection (e.g., “city”), alignment of comparable attributes (e.g., “population” of the target “cities”), contrast calculation (larger or smaller), as well as binary classification of positive and negative answers. The experimentation on HotpotQA illustrates that the proposed method achieves an average F1 score of 63.09%, outperforming the baseline with about 10% F1 scores. In addition, it performs better than a series of competitive models, including DecompRC, BERT.
“东京和纽约,哪个城市人口更多?”一般来说,为了回答这个问题,我们必须获得关于两个城市人口的先验知识,并相应地通过数字比较确定答案。使用机器阅读理解(MRC)来回答这样的问题已经成为一个热门的研究课题,这被称为比较问答任务(CQA)。在本文中,我们提出了一种新的神经CQA模型,该模型被训练来回答比较问题。该模型被设计为一个复杂的神经网络,它在一步一步的管道中执行推理,包括注意实体检测(例如,“城市”),可比较属性的校准(例如,目标“城市”的“人口”),对比计算(较大或较小),以及积极和消极答案的二进制分类。在HotpotQA上的实验表明,该方法的平均F1分数为63.09%,比基线的F1分数高出约10%。此外,它的性能优于一系列竞争模型,包括DecompRC, BERT。
{"title":"How to Answer Comparison Questions","authors":"Hongxuan Tang, Yu Hong, Xin Chen, Kaili Wu, Min Zhang","doi":"10.1109/IALP48816.2019.9037729","DOIUrl":"https://doi.org/10.1109/IALP48816.2019.9037729","url":null,"abstract":"“Which city has the larger population, Tokyo or New York?”. To answer the question, in general, we necessarily obtain the prior knowledge about the populations of both cities, and accordingly determine the answer by numeric comparison. Using Machine Reading Comprehension (MRC) to answer such a question has become a popular research topic, which is referred to as a task of Comparison Question Answering (CQA). In this paper, we propose a novel neural CQA model which is trained to answer comparison question. The model is designed as a sophisticated neural network which performs inference in a step-by-step pipeline, including the steps of attentive entity detection (e.g., “city”), alignment of comparable attributes (e.g., “population” of the target “cities”), contrast calculation (larger or smaller), as well as binary classification of positive and negative answers. The experimentation on HotpotQA illustrates that the proposed method achieves an average F1 score of 63.09%, outperforming the baseline with about 10% F1 scores. In addition, it performs better than a series of competitive models, including DecompRC, BERT.","PeriodicalId":208066,"journal":{"name":"2019 International Conference on Asian Language Processing (IALP)","volume":"110 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115758401","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
An Enhancement of Malay Social Media Text Normalization for Lexicon-Based Sentiment Analysis 马来语社交媒体文本规范化在基于词汇的情感分析中的改进
Pub Date : 2019-11-01 DOI: 10.1109/IALP48816.2019.9037700
Muhammad Fakhrur Razi Abu Bakar, N. Idris, Liyana Shuib
Nowadays, most Malaysians use social media such as Twitter to express their opinions toward any latest issues publicly. However, user individuality and creativity of language create huge volumes of noisy words which become unsuitable as dataset for any Natural Language Processing applications such as sentiment analysis due to the irregularity of the language featured. Thus, it is important to convert these noisy words into their standard forms. Currently, there are limited studies to normalize the noisy words for Malay language. Hence, the aim of this study is to propose an enhancement of Malay social media text normalization for lexicon-based sentiment analysis. This normalizer comprises six main modules: (1) advanced tokenization, (2) Malay/English token detection, (3) lexical rules, (4) noisy token replacement, (5) n-gram, and (6) detokenization. The evaluation has been conducted and the findings show that 83.55% achieved in Precision and 84.61% in Recall.
如今,大多数马来西亚人使用Twitter等社交媒体公开表达他们对任何最新问题的看法。然而,用户的个性和语言的创造性产生了大量的噪声词,由于语言特征的不规则性,这些词不适合作为情感分析等自然语言处理应用的数据集。因此,将这些嘈杂的单词转换成标准形式是很重要的。目前,对马来语中嘈杂词的规范化研究有限。因此,本研究的目的是为基于词汇的情感分析提出马来语社交媒体文本规范化的增强方法。该规范化程序包括六个主要模块:(1)高级标记化,(2)马来语/英语标记检测,(3)词法规则,(4)噪声标记替换,(5)n-gram,(6)去标记化。评估结果表明,准确率达到83.55%,召回率达到84.61%。
{"title":"An Enhancement of Malay Social Media Text Normalization for Lexicon-Based Sentiment Analysis","authors":"Muhammad Fakhrur Razi Abu Bakar, N. Idris, Liyana Shuib","doi":"10.1109/IALP48816.2019.9037700","DOIUrl":"https://doi.org/10.1109/IALP48816.2019.9037700","url":null,"abstract":"Nowadays, most Malaysians use social media such as Twitter to express their opinions toward any latest issues publicly. However, user individuality and creativity of language create huge volumes of noisy words which become unsuitable as dataset for any Natural Language Processing applications such as sentiment analysis due to the irregularity of the language featured. Thus, it is important to convert these noisy words into their standard forms. Currently, there are limited studies to normalize the noisy words for Malay language. Hence, the aim of this study is to propose an enhancement of Malay social media text normalization for lexicon-based sentiment analysis. This normalizer comprises six main modules: (1) advanced tokenization, (2) Malay/English token detection, (3) lexical rules, (4) noisy token replacement, (5) n-gram, and (6) detokenization. The evaluation has been conducted and the findings show that 83.55% achieved in Precision and 84.61% in Recall.","PeriodicalId":208066,"journal":{"name":"2019 International Conference on Asian Language Processing (IALP)","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116147733","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Acoustic Scene Classification Using Deep Convolutional Neural Network via Transfer Learning 基于迁移学习的深度卷积神经网络声学场景分类
Pub Date : 2019-11-01 DOI: 10.1109/IALP48816.2019.9037692
Min Ye, Hong Zhong, Xiao Song, Shilei Huang, Gang Cheng
We use deep convolutional neural network via transfer learning for Acoustic Scene Classification (ASC). For this purpose, a powerful and popular deep learning architecture — Residual Neural Network (Resnet) is adopted. Transfer learning is used to fine-tune the pre-trained Resnet model on the TUT Urban Acoustic Scenes 2018 dataset. Furthermore, the focal loss is used to improve overall performance. In order to reduce the chance of overfitting, data augmentation technique is applied based on mixup. Our best system has achieved an improvement of more than 10% in terms of class-wise accuracy with respect to the Detection and classification of acoustic scenes and events (DCASE) 2018 baseline system on the TUT Urban Acoustic Scenes 2018 dataset.
将基于迁移学习的深度卷积神经网络用于声学场景分类。为此,采用了一种强大而流行的深度学习架构——残余神经网络(Resnet)。迁移学习用于在TUT城市声学场景2018数据集上微调预训练的Resnet模型。此外,焦损被用来提高整体性能。为了减少过拟合的可能性,采用了基于混合的数据增强技术。在TUT城市声学场景2018数据集上,我们最好的系统在声学场景和事件的检测和分类(DCASE) 2018基线系统方面的分类精度提高了10%以上。
{"title":"Acoustic Scene Classification Using Deep Convolutional Neural Network via Transfer Learning","authors":"Min Ye, Hong Zhong, Xiao Song, Shilei Huang, Gang Cheng","doi":"10.1109/IALP48816.2019.9037692","DOIUrl":"https://doi.org/10.1109/IALP48816.2019.9037692","url":null,"abstract":"We use deep convolutional neural network via transfer learning for Acoustic Scene Classification (ASC). For this purpose, a powerful and popular deep learning architecture — Residual Neural Network (Resnet) is adopted. Transfer learning is used to fine-tune the pre-trained Resnet model on the TUT Urban Acoustic Scenes 2018 dataset. Furthermore, the focal loss is used to improve overall performance. In order to reduce the chance of overfitting, data augmentation technique is applied based on mixup. Our best system has achieved an improvement of more than 10% in terms of class-wise accuracy with respect to the Detection and classification of acoustic scenes and events (DCASE) 2018 baseline system on the TUT Urban Acoustic Scenes 2018 dataset.","PeriodicalId":208066,"journal":{"name":"2019 International Conference on Asian Language Processing (IALP)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121324613","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Sinhala and Tamil Speech Intent Identification From English Phoneme Based ASR 基于英语音素ASR的僧伽罗语和泰米尔语语音意图识别
Pub Date : 2019-11-01 DOI: 10.1109/IALP48816.2019.9037702
Yohan Karunanayake, Uthayasanker Thayasivam, Surangika Ranathunga
Today we can find many use cases for content-based speech classification. These include speech topic identification and spoken command recognition. Automatic Speech Recognition (ASR) sits underneath all of these applications to convert speech into textual format. However, creating an ASR system for a language is a resource-consuming task. Even though there are more than 6000 languages, all of these speech-related applications are limited to the most well-known languages such as English, because of the availability of data. There is some past research that looked into classifying speech while addressing the data scarcity. However, all of these methods have their own limitations. In this paper, we present an English language phoneme based speech intent classification methodology for Sinhala and Tamil languages. We use a pre-trained English ASR model to generate phoneme probability features and use them to identify intents of utterances expressed in Sinhala and Tamil, for which a rather small speech dataset is available. The experiment results show that the proposed method can have more than 80% accuracy for a 0.5-hour limited speech dataset in both languages.
今天,我们可以找到许多基于内容的语音分类的用例。这包括语音主题识别和语音命令识别。自动语音识别(ASR)位于所有这些应用程序的下面,将语音转换为文本格式。然而,为一种语言创建ASR系统是一项消耗资源的任务。尽管有超过6000种语言,但由于数据的可用性,所有这些与语音相关的应用程序都仅限于最知名的语言,如英语。过去有一些研究着眼于对语音进行分类,同时解决数据稀缺问题。然而,所有这些方法都有其自身的局限性。本文提出了一种基于英语音位的僧伽罗语和泰米尔语语音意图分类方法。我们使用预训练的英语ASR模型来生成音素概率特征,并使用它们来识别僧伽罗语和泰米尔语表达的话语意图,这两种语言的语音数据集相当小。实验结果表明,对于两种语言的0.5小时限定语音数据集,该方法的准确率均在80%以上。
{"title":"Sinhala and Tamil Speech Intent Identification From English Phoneme Based ASR","authors":"Yohan Karunanayake, Uthayasanker Thayasivam, Surangika Ranathunga","doi":"10.1109/IALP48816.2019.9037702","DOIUrl":"https://doi.org/10.1109/IALP48816.2019.9037702","url":null,"abstract":"Today we can find many use cases for content-based speech classification. These include speech topic identification and spoken command recognition. Automatic Speech Recognition (ASR) sits underneath all of these applications to convert speech into textual format. However, creating an ASR system for a language is a resource-consuming task. Even though there are more than 6000 languages, all of these speech-related applications are limited to the most well-known languages such as English, because of the availability of data. There is some past research that looked into classifying speech while addressing the data scarcity. However, all of these methods have their own limitations. In this paper, we present an English language phoneme based speech intent classification methodology for Sinhala and Tamil languages. We use a pre-trained English ASR model to generate phoneme probability features and use them to identify intents of utterances expressed in Sinhala and Tamil, for which a rather small speech dataset is available. The experiment results show that the proposed method can have more than 80% accuracy for a 0.5-hour limited speech dataset in both languages.","PeriodicalId":208066,"journal":{"name":"2019 International Conference on Asian Language Processing (IALP)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121422077","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Design and Implementation of Burmese Speech Synthesis System Based on HMM-DNN 基于HMM-DNN的缅甸语语音合成系统的设计与实现
Pub Date : 2019-11-01 DOI: 10.1109/IALP48816.2019.9037731
Mengyuan Liu, Jian Yang
The research and application of speech synthesis in Chinese and English are widely used. However, most nonuniversal languages have relatively few electronic language resources, and speech synthesis research is lagging behind. Burmese is a type of alphabetic writing, and Burmese belongs to Tibetan-Burmese branch of the Sino-Tibetan language. In order to develop the Burmese speech synthesis application system, this paper studies the Burmese speech waveform synthesis method, designs and implements a HMM-based Burmese speech synthesis baseline system, and based on this, introduces a deep neural network (DNN) to replace the decision tree model of HMM speech synthesis system, thereby improving the acoustic model to improve the quality of speech synthesis. The experimental results show that the baseline system is feasible, and the introduction of DNN speech synthesis system can effectively improve the quality of speech synthesis.
语音合成的研究和应用在汉语和英语中得到了广泛的应用。然而,大多数非通用语言的电子语言资源相对较少,语音合成研究相对滞后。缅甸语是一种字母文字,缅甸语属于汉藏语系的藏缅语分支。为了开发缅甸语语音合成应用系统,本文研究了缅甸语语音波形合成方法,设计并实现了基于HMM的缅甸语语音合成基线系统,并在此基础上引入深度神经网络(DNN)取代HMM语音合成系统的决策树模型,从而改进声学模型,提高语音合成质量。实验结果表明,基线系统是可行的,DNN语音合成系统的引入可以有效提高语音合成的质量。
{"title":"Design and Implementation of Burmese Speech Synthesis System Based on HMM-DNN","authors":"Mengyuan Liu, Jian Yang","doi":"10.1109/IALP48816.2019.9037731","DOIUrl":"https://doi.org/10.1109/IALP48816.2019.9037731","url":null,"abstract":"The research and application of speech synthesis in Chinese and English are widely used. However, most nonuniversal languages have relatively few electronic language resources, and speech synthesis research is lagging behind. Burmese is a type of alphabetic writing, and Burmese belongs to Tibetan-Burmese branch of the Sino-Tibetan language. In order to develop the Burmese speech synthesis application system, this paper studies the Burmese speech waveform synthesis method, designs and implements a HMM-based Burmese speech synthesis baseline system, and based on this, introduces a deep neural network (DNN) to replace the decision tree model of HMM speech synthesis system, thereby improving the acoustic model to improve the quality of speech synthesis. The experimental results show that the baseline system is feasible, and the introduction of DNN speech synthesis system can effectively improve the quality of speech synthesis.","PeriodicalId":208066,"journal":{"name":"2019 International Conference on Asian Language Processing (IALP)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128427279","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Learning Deep Matching-Aware Network for Text Recommendation using Clickthrough Data 使用点击数据学习文本推荐的深度匹配感知网络
Pub Date : 2019-11-01 DOI: 10.1109/IALP48816.2019.9037682
Haonan Liu, Nankai Lin, Zitao Chen, Ke Li, Sheng-yi Jiang
With the trend of information globalization, the volume of text information is exploding, which results in the information overload problem. Text recommendation system has shown to be a valuable tool to help users in such situations of information overload. In general, most researchers define text recommendation as a static problem, ignoring sequential information. In this paper, we propose a text recommendation framework with matching-aware interest extractor and dynamic interest extractor. We apply the Attention-based Long Short-Term Memory Network (LSTM) to model a user’ s dynamic interest. Besides, we model a user’ s static interest with the idea of semantic matching. We integrate dynamic interest and static interest of users’ and decide whether to recommend a text. We also propose a reasonable method to construct a text recommendation dataset with clickthrough data from CCIR 2018 shared task Personal Recommendation. We test our model and other baseline models on the dataset. The experiment shows our model outperforms all the baseline models and a state-of-the-art model, and the Fl-score of our model reaches 0.76.
随着信息全球化的趋势,文本信息量呈爆炸式增长,导致了信息过载问题。文本推荐系统已被证明是一种有价值的工具,可以帮助用户在这种信息过载的情况下。一般来说,大多数研究者将文本推荐定义为一个静态问题,忽略了顺序信息。本文提出了一种具有匹配感知兴趣提取器和动态兴趣提取器的文本推荐框架。我们应用基于注意的长短期记忆网络(LSTM)来模拟用户的动态兴趣。此外,我们利用语义匹配的思想对用户的静态兴趣进行建模。我们综合用户的动态兴趣和静态兴趣来决定是否推荐文本。我们还提出了一种合理的方法,利用CCIR 2018共享任务个人推荐的点击率数据构建文本推荐数据集。我们在数据集上测试我们的模型和其他基线模型。实验表明,我们的模型优于所有的基线模型和最先进的模型,我们的模型的fl得分达到0.76。
{"title":"Learning Deep Matching-Aware Network for Text Recommendation using Clickthrough Data","authors":"Haonan Liu, Nankai Lin, Zitao Chen, Ke Li, Sheng-yi Jiang","doi":"10.1109/IALP48816.2019.9037682","DOIUrl":"https://doi.org/10.1109/IALP48816.2019.9037682","url":null,"abstract":"With the trend of information globalization, the volume of text information is exploding, which results in the information overload problem. Text recommendation system has shown to be a valuable tool to help users in such situations of information overload. In general, most researchers define text recommendation as a static problem, ignoring sequential information. In this paper, we propose a text recommendation framework with matching-aware interest extractor and dynamic interest extractor. We apply the Attention-based Long Short-Term Memory Network (LSTM) to model a user’ s dynamic interest. Besides, we model a user’ s static interest with the idea of semantic matching. We integrate dynamic interest and static interest of users’ and decide whether to recommend a text. We also propose a reasonable method to construct a text recommendation dataset with clickthrough data from CCIR 2018 shared task Personal Recommendation. We test our model and other baseline models on the dataset. The experiment shows our model outperforms all the baseline models and a state-of-the-art model, and the Fl-score of our model reaches 0.76.","PeriodicalId":208066,"journal":{"name":"2019 International Conference on Asian Language Processing (IALP)","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126418842","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
combination of Semantic Relatedness with Supervised Method for Word Sense Disambiguation 语义关联与监督相结合的词义消歧方法
Pub Date : 2019-11-01 DOI: 10.1109/IALP48816.2019.9037717
Qiaoli Zhou, Yuguang Meng
We present a semi-supervised learning method at efficiently exploits semantic relatedness in order to incorporate sense knowledge into a word sense disambiguation model and to leverage system performance. We have presented sense relativeness algorithms which combine neural model learned from a generic embedding function for variable length contexts of target words on a POS-labeled text corpus, with sense-labeled data in the form of example sentences. This paper investigates the way of incorporating semantic relatedness in a word sense disambiguation setting and evaluates the method on some SensEval/SemEval lexical sample tasks. The obtained results show that such representations consistently improve the accuracy of the selective supervised WSD system.
我们提出了一种半监督学习方法,有效地利用语义相关性,将语义知识整合到词义消歧模型中,并利用系统性能。我们提出了一种语义相关性算法,该算法将从一个通用嵌入函数中学习到的神经模型与poss标记文本语料库上目标词的变长上下文相结合,并以例句的形式进行语义标记数据。本文研究了在词义消歧设置中引入语义关联的方法,并在一些SensEval/SemEval词汇样本任务中对该方法进行了评价。得到的结果表明,这种表示一致地提高了选择性监督WSD系统的精度。
{"title":"combination of Semantic Relatedness with Supervised Method for Word Sense Disambiguation","authors":"Qiaoli Zhou, Yuguang Meng","doi":"10.1109/IALP48816.2019.9037717","DOIUrl":"https://doi.org/10.1109/IALP48816.2019.9037717","url":null,"abstract":"We present a semi-supervised learning method at efficiently exploits semantic relatedness in order to incorporate sense knowledge into a word sense disambiguation model and to leverage system performance. We have presented sense relativeness algorithms which combine neural model learned from a generic embedding function for variable length contexts of target words on a POS-labeled text corpus, with sense-labeled data in the form of example sentences. This paper investigates the way of incorporating semantic relatedness in a word sense disambiguation setting and evaluates the method on some SensEval/SemEval lexical sample tasks. The obtained results show that such representations consistently improve the accuracy of the selective supervised WSD system.","PeriodicalId":208066,"journal":{"name":"2019 International Conference on Asian Language Processing (IALP)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133115486","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Tibetan word segmentation method based on CNN-BiLSTM-CRF model 基于CNN-BiLSTM-CRF模型的藏文分词方法
Pub Date : 2018-11-01 DOI: 10.1109/IALP48816.2019.9037661
Lili Wang, Hongwu Yang, Xiaotian Xing, Yajing Yan
We propose a Tibetan word segmentation method based on CNN-BiLSTM-CRF model that merely uses the characters of sentence as the input so that the method does not need large-scale corpus resources and manual features for training. Firstly, we use convolution neural network to train character vectors. Then the character vectors are searched through the character lookup table to form a matrix C by stacking searched results. Then the convolution operation between the matrix C and multiple filter matrices is carried out to obtain the character-level features of each Tibetan word by maximizing the pooling. We input the character vector into the BiLSTM-CRF model, which is suitable for Tibetan word segmentation through the highway network, for getting a Tibetan word segmentation model that is optimized by using the character vector and CRF model. For Tibetan language with rich morphology, fewer parameters and faster training time make this model better than BiLSTM-CRF model in the performance of character level. The experimental results show that character input is sufficient for language modeling. The robustness of Tibetan word segmentation is improved by the model that can achieves 95.17% of the F value.
本文提出了一种基于CNN-BiLSTM-CRF模型的藏文分词方法,该方法仅使用句子的字符作为输入,不需要大规模的语料库资源和人工特征进行训练。首先利用卷积神经网络对特征向量进行训练。然后通过字符查找表搜索字符向量,将搜索结果叠加形成矩阵C。然后对矩阵C与多个滤波矩阵进行卷积运算,通过池化最大化的方法获得每个藏文词的字符级特征。我们将特征向量输入到适用于公路网络藏文分词的BiLSTM-CRF模型中,得到一个结合特征向量和CRF模型进行优化的藏文分词模型。对于形态丰富的藏语,该模型参数更少,训练时间更快,在字符水平上优于BiLSTM-CRF模型。实验结果表明,字符输入对语言建模是足够的。该模型提高了藏文分词的鲁棒性,可达到F值的95.17%。
{"title":"Tibetan word segmentation method based on CNN-BiLSTM-CRF model","authors":"Lili Wang, Hongwu Yang, Xiaotian Xing, Yajing Yan","doi":"10.1109/IALP48816.2019.9037661","DOIUrl":"https://doi.org/10.1109/IALP48816.2019.9037661","url":null,"abstract":"We propose a Tibetan word segmentation method based on CNN-BiLSTM-CRF model that merely uses the characters of sentence as the input so that the method does not need large-scale corpus resources and manual features for training. Firstly, we use convolution neural network to train character vectors. Then the character vectors are searched through the character lookup table to form a matrix C by stacking searched results. Then the convolution operation between the matrix C and multiple filter matrices is carried out to obtain the character-level features of each Tibetan word by maximizing the pooling. We input the character vector into the BiLSTM-CRF model, which is suitable for Tibetan word segmentation through the highway network, for getting a Tibetan word segmentation model that is optimized by using the character vector and CRF model. For Tibetan language with rich morphology, fewer parameters and faster training time make this model better than BiLSTM-CRF model in the performance of character level. The experimental results show that character input is sufficient for language modeling. The robustness of Tibetan word segmentation is improved by the model that can achieves 95.17% of the F value.","PeriodicalId":208066,"journal":{"name":"2019 International Conference on Asian Language Processing (IALP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127994165","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
2019 International Conference on Asian Language Processing (IALP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1