首页 > 最新文献

2019 International Conference on Asian Language Processing (IALP)最新文献

英文 中文
Using Mention Segmentation to Improve Event Detection with Multi-head Attention 基于提及分割改进多头注意事件检测
Pub Date : 2019-11-01 DOI: 10.1109/IALP48816.2019.9037683
Jiali Chen, Yu Hong, Jingli Zhang, Jianmin Yao
Sentence-level event detection (ED) is a task of detecting words that describe specific types of events, including the subtasks of trigger word identification and event type classification. Previous work straightforwardly inputs a sentence into neural classification models and analyzes deep semantics of words in the sentence one by one. Relying on the semantics, probabilities of event classes can be predicted for each word, including the carefully defined ACE event classes and a “N/A” class(i.e., non-trigger word). The models achieve remarkable successes nowadays. However, our findings show that a natural sentence may posses more than one trigger word and thus entail different types of events. In particular, the closely related information of each event only lies in a unique sentence segment but has nothing to do with other segments. In order to reduce negative influences from noises in other segments, we propose to perform semantics learning for event detection only in the scope of segment instead of the whole sentence. Accordingly, we develop a novel ED method which integrates sentence segmentation into the neural event classification architecture. Bidirectional Long Short-Term Memory (Bi-LSTM) with multi-head attention is used as the classification model. Sentence segmentation is boiled down to a sequence labeling problem, where BERT is used. We combine embeddings, and use them as the input of the neural classification model. The experimental results show that the performance of our method reaches 76.8% and 74.2% $F_{1}-$scores for trigger identification and event type classification, which outperforms the state-of-the-art.
句子级事件检测是一项检测描述特定事件类型的词的任务,包括触发词识别和事件类型分类的子任务。以前的工作是直接将一个句子输入到神经分类模型中,逐个分析句子中单词的深层语义。依靠语义,可以预测每个单词的事件类概率,包括仔细定义的ACE事件类和“N/ a”类(即。(非触发词)。这些模型如今取得了显著的成功。然而,我们的研究结果表明,一个自然的句子可能有多个触发词,从而导致不同类型的事件。特别是,每个事件密切相关的信息只存在于一个独特的句段中,与其他句段无关。为了减少其他语段中噪声的负面影响,我们建议仅在语段范围内而不是在整个句子范围内进行语义学习以进行事件检测。因此,我们开发了一种新的ED方法,该方法将句子分词与神经事件分类体系结构相结合。采用具有多头注意的双向长短期记忆作为分类模型。句子切分被归结为一个序列标记问题,其中使用了BERT。我们结合嵌入,并使用它们作为神经分类模型的输入。实验结果表明,该方法在触发识别和事件类型分类上的得分分别达到76.8%和74.2%。
{"title":"Using Mention Segmentation to Improve Event Detection with Multi-head Attention","authors":"Jiali Chen, Yu Hong, Jingli Zhang, Jianmin Yao","doi":"10.1109/IALP48816.2019.9037683","DOIUrl":"https://doi.org/10.1109/IALP48816.2019.9037683","url":null,"abstract":"Sentence-level event detection (ED) is a task of detecting words that describe specific types of events, including the subtasks of trigger word identification and event type classification. Previous work straightforwardly inputs a sentence into neural classification models and analyzes deep semantics of words in the sentence one by one. Relying on the semantics, probabilities of event classes can be predicted for each word, including the carefully defined ACE event classes and a “N/A” class(i.e., non-trigger word). The models achieve remarkable successes nowadays. However, our findings show that a natural sentence may posses more than one trigger word and thus entail different types of events. In particular, the closely related information of each event only lies in a unique sentence segment but has nothing to do with other segments. In order to reduce negative influences from noises in other segments, we propose to perform semantics learning for event detection only in the scope of segment instead of the whole sentence. Accordingly, we develop a novel ED method which integrates sentence segmentation into the neural event classification architecture. Bidirectional Long Short-Term Memory (Bi-LSTM) with multi-head attention is used as the classification model. Sentence segmentation is boiled down to a sequence labeling problem, where BERT is used. We combine embeddings, and use them as the input of the neural classification model. The experimental results show that the performance of our method reaches 76.8% and 74.2% $F_{1}-$scores for trigger identification and event type classification, which outperforms the state-of-the-art.","PeriodicalId":208066,"journal":{"name":"2019 International Conference on Asian Language Processing (IALP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129631229","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic Extraction and Quantitative Evaluation of the Character Relationship Networks from Children’s Literature works 儿童文学作品中人物关系网络的自动提取与定量评价
Pub Date : 2019-11-01 DOI: 10.1109/IALP48816.2019.9037669
Kun Ma, Lijiao Yang
To automate the graded reading task, we urgently need to extract and calculate the important index of the complexity of the relationship between the characters affecting the plot complexity of narrative literature. In order to realize this purpose, this paper describes a computational method for automatic analysis of the virtual social network from children’s literature works. We selected the required bibliography for primary school students recommended by the Ministry of Education, then automatically extract the characters of the novel by CRF, and constructs the character network based on the co-occurrence relationship. The statistical analysis method of complex network provides a quantitative basis for distinguishing the complexity of characters’ relationships in different texts. The results show that the structural characteristics of character interaction networks are similar to those of small world networks, and the selected network measurement indexes are significantly related to the complexity of text characters. Finally, we achieved effectively evaluating and predicting the complexity of the social networks from more extensive literature works some classical regression model based on machine learning.
为了实现分级阅读任务的自动化,我们迫切需要提取和计算影响叙事文学情节复杂性的人物关系复杂性这一重要指标。为了实现这一目的,本文描述了一种自动分析儿童文学作品虚拟社会网络的计算方法。选取教育部推荐的小学生所需书目,利用CRF自动提取小说人物,并基于共现关系构建人物网络。复杂网络的统计分析方法为区分不同文本中人物关系的复杂性提供了定量依据。结果表明,字符交互网络的结构特征与小世界网络相似,所选择的网络度量指标与文本字符的复杂程度有显著相关。最后,基于机器学习的经典回归模型在更广泛的文献中实现了对社会网络复杂性的有效评估和预测。
{"title":"Automatic Extraction and Quantitative Evaluation of the Character Relationship Networks from Children’s Literature works","authors":"Kun Ma, Lijiao Yang","doi":"10.1109/IALP48816.2019.9037669","DOIUrl":"https://doi.org/10.1109/IALP48816.2019.9037669","url":null,"abstract":"To automate the graded reading task, we urgently need to extract and calculate the important index of the complexity of the relationship between the characters affecting the plot complexity of narrative literature. In order to realize this purpose, this paper describes a computational method for automatic analysis of the virtual social network from children’s literature works. We selected the required bibliography for primary school students recommended by the Ministry of Education, then automatically extract the characters of the novel by CRF, and constructs the character network based on the co-occurrence relationship. The statistical analysis method of complex network provides a quantitative basis for distinguishing the complexity of characters’ relationships in different texts. The results show that the structural characteristics of character interaction networks are similar to those of small world networks, and the selected network measurement indexes are significantly related to the complexity of text characters. Finally, we achieved effectively evaluating and predicting the complexity of the social networks from more extensive literature works some classical regression model based on machine learning.","PeriodicalId":208066,"journal":{"name":"2019 International Conference on Asian Language Processing (IALP)","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120907917","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
An experimental Tagalog Finite State Automata spellchecker with Levenshtein edit-distance feature 具有Levenshtein编辑距离特征的实验性他加禄语有限状态自动机拼写检查器
Pub Date : 2019-11-01 DOI: 10.1109/IALP48816.2019.9037687
Joseph Marvin Imperial, C. G. V. Ya-On, Jennifer C. Ureta
In this paper, we present an experimental development of a spell checker for the Tagalog language using a set of word list with 300 random root words and three inflected forms as training data and a two-layered architecture of combined Deterministic Finite Automaton (DFA) with Levenshtein edit-distance. A DFA is used to process strings to identify if it belongs to a certain language via the binary result of accept or reject. The Levenshtein edit-distance of two strings is the number (k) of deletions, alterations, insertions between two sequences of characters. From the sample trained wordlist, results show that a value of 1 for the edit-distance (k) can be effective in spelling Tagalog sentences. Any value greater than 1 can cause suggestion of words even if the spelling of words is correct due to selective and prominent usage of certain characters in the Tagalog language like a, n, g, t, s, l.
在本文中,我们使用一组包含300个随机词根词和三种屈变形式的单词列表作为训练数据,并结合确定性有限自动机(DFA)和Levenshtein编辑距离的两层结构,提出了一种针对他加洛语的拼写检查器的实验开发。DFA用于处理字符串,通过接受或拒绝的二进制结果来识别它是否属于某种语言。两个字符串的Levenshtein编辑距离是两个字符序列之间删除、修改、插入的次数(k)。从样本训练的单词列表中,结果表明编辑距离(k)的值为1可以有效地拼写他加禄语句子。任何大于1的值都可能导致单词的提示,即使单词的拼写是正确的,因为在他加禄语中有选择地突出使用某些字符,如a、n、g、t、s、l。
{"title":"An experimental Tagalog Finite State Automata spellchecker with Levenshtein edit-distance feature","authors":"Joseph Marvin Imperial, C. G. V. Ya-On, Jennifer C. Ureta","doi":"10.1109/IALP48816.2019.9037687","DOIUrl":"https://doi.org/10.1109/IALP48816.2019.9037687","url":null,"abstract":"In this paper, we present an experimental development of a spell checker for the Tagalog language using a set of word list with 300 random root words and three inflected forms as training data and a two-layered architecture of combined Deterministic Finite Automaton (DFA) with Levenshtein edit-distance. A DFA is used to process strings to identify if it belongs to a certain language via the binary result of accept or reject. The Levenshtein edit-distance of two strings is the number (k) of deletions, alterations, insertions between two sequences of characters. From the sample trained wordlist, results show that a value of 1 for the edit-distance (k) can be effective in spelling Tagalog sentences. Any value greater than 1 can cause suggestion of words even if the spelling of words is correct due to selective and prominent usage of certain characters in the Tagalog language like a, n, g, t, s, l.","PeriodicalId":208066,"journal":{"name":"2019 International Conference on Asian Language Processing (IALP)","volume":"2 2","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121004233","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Acoustic Cues for Sensory Modality and Referential Vagueness of Personal Pronouns: Me or Not? You or Not? 人称代词感官情态和指称模糊的声学线索:我还是不是?你还是不是?
Pub Date : 2019-11-01 DOI: 10.1109/IALP48816.2019.9037676
Aibao Zhou, Yanbing Hu, Xiaoyong Lu, Tao Pan
Personal pronouns are of great pragmatic significance, and among their numerous functions, referential vagueness is the key to the exploration of the self. In Experiment 1, the relationship of first-, second-, and third personal pronouns with the self was discussed in acoustic condition, with gender consistency, pronoun type, and person type as independent variables. Experiment 2 records the frequency of the subjects on the SR button and the NSR button, and statistical analysis was performed on the SR button reaction. This study finds that the same pronouns show different self-cognitive processing under the different visual and acoustic stimuli, and the results support the dynamics of personal pronouns and possessive pronouns and self relationships.
人称代词具有重要的语用意义,在其众多功能中,指称模糊是探索自我的关键。实验1以性别一致性、代词类型和人类型为自变量,探讨声学条件下第一、第二、第三人称代词与自我的关系。实验2记录被试按SR按钮和按NSR按钮的频率,并对SR按钮反应进行统计分析。本研究发现,同一人称代词在不同视觉和听觉刺激下表现出不同的自我认知加工,这一结果支持人称代词和所有格代词与自我关系的动态关系。
{"title":"Acoustic Cues for Sensory Modality and Referential Vagueness of Personal Pronouns: Me or Not? You or Not?","authors":"Aibao Zhou, Yanbing Hu, Xiaoyong Lu, Tao Pan","doi":"10.1109/IALP48816.2019.9037676","DOIUrl":"https://doi.org/10.1109/IALP48816.2019.9037676","url":null,"abstract":"Personal pronouns are of great pragmatic significance, and among their numerous functions, referential vagueness is the key to the exploration of the self. In Experiment 1, the relationship of first-, second-, and third personal pronouns with the self was discussed in acoustic condition, with gender consistency, pronoun type, and person type as independent variables. Experiment 2 records the frequency of the subjects on the SR button and the NSR button, and statistical analysis was performed on the SR button reaction. This study finds that the same pronouns show different self-cognitive processing under the different visual and acoustic stimuli, and the results support the dynamics of personal pronouns and possessive pronouns and self relationships.","PeriodicalId":208066,"journal":{"name":"2019 International Conference on Asian Language Processing (IALP)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122703299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Measurement Method of Ancient Poetry Difficulty for Adaptive Testing 古诗难度自适应测试方法研究
Pub Date : 2019-11-01 DOI: 10.1109/IALP48816.2019.9037665
Huiping Wang, B. Wang
As traditional Chinese culture education grows fast, adaptive testing for ancient poetry seems promising. The prerequisite of adaptive learning is question bank, while the quality of question bank depends on the rationality of question difficulty. The paper proposes a method that measures ancient poetry difficulty for objective questions and builds a measurement system. The method includes five steps: (1) Identify the verses corresponding to the question. (2) Get four indexes of the corresponding verses: search volume, correctly answered frequency, sentence length and grade of the textbook which includes the verses.(3) Use analytic hierarchy process to index system for weight assignment. (4) Compute the weighted sum of the four indexes as the measurement of difficulty. (5) Classify the question bank according to the calculated difficulty. Experimental results showed the effectiveness of this measurement method, which thereby can be used in various adaptive testing of ancient poetry.
随着中国传统文化教育的快速发展,古诗适应性测试似乎大有发展前景。适应性学习的前提是题库,而题库的质量取决于题目难度的合理性。本文提出了一种对客观问题进行古诗难度测量的方法,并构建了测量体系。该方法包括五个步骤:(1)找出与问题对应的经文。(2)得到相应诗句的四个指标:搜索量、正确回答频率、句子长度和包含诗句的教科书等级。(3)利用层次分析法建立指标体系,进行权重分配。(4)计算四项指标的加权和作为难度的度量。(5)根据计算出的难度对题库进行分类。实验结果表明了该测量方法的有效性,可用于古诗的各种适应性测试。
{"title":"An Measurement Method of Ancient Poetry Difficulty for Adaptive Testing","authors":"Huiping Wang, B. Wang","doi":"10.1109/IALP48816.2019.9037665","DOIUrl":"https://doi.org/10.1109/IALP48816.2019.9037665","url":null,"abstract":"As traditional Chinese culture education grows fast, adaptive testing for ancient poetry seems promising. The prerequisite of adaptive learning is question bank, while the quality of question bank depends on the rationality of question difficulty. The paper proposes a method that measures ancient poetry difficulty for objective questions and builds a measurement system. The method includes five steps: (1) Identify the verses corresponding to the question. (2) Get four indexes of the corresponding verses: search volume, correctly answered frequency, sentence length and grade of the textbook which includes the verses.(3) Use analytic hierarchy process to index system for weight assignment. (4) Compute the weighted sum of the four indexes as the measurement of difficulty. (5) Classify the question bank according to the calculated difficulty. Experimental results showed the effectiveness of this measurement method, which thereby can be used in various adaptive testing of ancient poetry.","PeriodicalId":208066,"journal":{"name":"2019 International Conference on Asian Language Processing (IALP)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115462910","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic Recognition of Allusions in Tang Poetry Based on BERT 基于BERT的唐诗典故自动识别
Pub Date : 2019-11-01 DOI: 10.1109/IALP48816.2019.9037679
Xuemei Tang, Shichen Liang, Jianyu Zheng, Renfen Hu, Zhiying Liu
In this paper, we propose an automated method for recognize allusions in Tang poetry. The representation of text is trained by BERT pre-trained by The SiKuQuanShu. The TOP-20 candidate allusions have the highest semantic similarity to the original sentence. Then update the ranking of candidate allusions by rule-based ranking algorithm. In the final experimental results, the precision of the correct allusion same as the final ranking TOP-I reached 63.74%, the precision of the correct allusion appears in the final ranking TOP-3 reached 70.66%, and the precision of the correct allusion appears in the final ranking TOP-5 reached 74.82%.
本文提出了一种自动识别唐诗典故的方法。文本的表示由The SiKuQuanShu预训练的BERT进行训练。前20个候选典故与原句的语义相似度最高。然后通过基于规则的排序算法更新候选典故的排序。在最终的实验结果中,与最终排名top - 1相同的正确典故的精度达到63.74%,最终排名TOP-3的正确典故出现的精度达到70.66%,最终排名TOP-5的正确典故出现的精度达到74.82%。
{"title":"Automatic Recognition of Allusions in Tang Poetry Based on BERT","authors":"Xuemei Tang, Shichen Liang, Jianyu Zheng, Renfen Hu, Zhiying Liu","doi":"10.1109/IALP48816.2019.9037679","DOIUrl":"https://doi.org/10.1109/IALP48816.2019.9037679","url":null,"abstract":"In this paper, we propose an automated method for recognize allusions in Tang poetry. The representation of text is trained by BERT pre-trained by The SiKuQuanShu. The TOP-20 candidate allusions have the highest semantic similarity to the original sentence. Then update the ranking of candidate allusions by rule-based ranking algorithm. In the final experimental results, the precision of the correct allusion same as the final ranking TOP-I reached 63.74%, the precision of the correct allusion appears in the final ranking TOP-3 reached 70.66%, and the precision of the correct allusion appears in the final ranking TOP-5 reached 74.82%.","PeriodicalId":208066,"journal":{"name":"2019 International Conference on Asian Language Processing (IALP)","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131524911","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Improving Mandarin Prosody Boundary Detection by Using Phonetic Information and Deep LSTM Model 基于语音信息和深度LSTM模型改进汉语韵律边界检测
Pub Date : 2019-11-01 DOI: 10.1109/IALP48816.2019.9037697
Ju Lin, Zhuanzhuan Ji, Wenwei Dong, Yanlu Xie, Jinsong Zhang
Automatic prosodic boundary detection is useful for automatic speech processing, such as automatic speech recognition (ASR) and speech synthesis. In this paper, we propose two techniques to improve the boundary detection performance. First, in addition to prosody features (e.g, pitch, duration and energy), phonetic information (word/articulatory information) is integrated into the framework of prosodic boundary detection. We compared two forms of phonetic information: word form and articulatory form. Moreover, boundary detection can be regarded as a sequence labeling task. A deep Long Short-Term Memory (LSTM) is adopted for this task, which replaces the traditional Deep Neural Networks (DNN) model. The experimental results showed that the boundary detection performance can be improved by the additional phonetic information, with relative 5.9% (word form) and 9.8% (articulatory form) improvements respectively in contrast with the system that only used prosody features modeled. The articulatory information and prosody features with deep LSTM achieved the best result, with further performance enhancement from 76.35% to 77.85% (relative 6.3%) compared with that modeled by DNN.
自动韵律边界检测对于自动语音识别和语音合成等自动语音处理具有重要意义。在本文中,我们提出了两种技术来提高边界检测性能。首先,在韵律特征(如音高、音长、能量)之外,将语音信息(词/发音信息)整合到韵律边界检测框架中。我们比较了两种形式的语音信息:词形和发音形式。此外,边界检测可以看作是一个序列标记任务。该任务采用深度长短期记忆(LSTM)模型来代替传统的深度神经网络(DNN)模型。实验结果表明,与仅使用韵律特征建模的系统相比,添加语音信息可以提高边界检测性能,分别提高5.9%(词形)和9.8%(发音形式)。深度LSTM在发音信息和韵律特征上取得了最好的效果,与DNN模型相比,性能进一步提高,从76.35%提高到77.85%(相对6.3%)。
{"title":"Improving Mandarin Prosody Boundary Detection by Using Phonetic Information and Deep LSTM Model","authors":"Ju Lin, Zhuanzhuan Ji, Wenwei Dong, Yanlu Xie, Jinsong Zhang","doi":"10.1109/IALP48816.2019.9037697","DOIUrl":"https://doi.org/10.1109/IALP48816.2019.9037697","url":null,"abstract":"Automatic prosodic boundary detection is useful for automatic speech processing, such as automatic speech recognition (ASR) and speech synthesis. In this paper, we propose two techniques to improve the boundary detection performance. First, in addition to prosody features (e.g, pitch, duration and energy), phonetic information (word/articulatory information) is integrated into the framework of prosodic boundary detection. We compared two forms of phonetic information: word form and articulatory form. Moreover, boundary detection can be regarded as a sequence labeling task. A deep Long Short-Term Memory (LSTM) is adopted for this task, which replaces the traditional Deep Neural Networks (DNN) model. The experimental results showed that the boundary detection performance can be improved by the additional phonetic information, with relative 5.9% (word form) and 9.8% (articulatory form) improvements respectively in contrast with the system that only used prosody features modeled. The articulatory information and prosody features with deep LSTM achieved the best result, with further performance enhancement from 76.35% to 77.85% (relative 6.3%) compared with that modeled by DNN.","PeriodicalId":208066,"journal":{"name":"2019 International Conference on Asian Language Processing (IALP)","volume":"688 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132184755","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Integrating Knowledge in End-to-End Automatic Speech Recognition for Mandarin-English Code-Switching 集成知识在端到端自动语音识别中的应用
Pub Date : 2019-11-01 DOI: 10.1109/IALP48816.2019.9037688
Chia-Yu Li, Ngoc Thang Vu
Code-Switching (CS) is a common linguistic phenomenon in multilingual communities that consists of switching between languages while speaking. This paper presents our investigations on end-to-end speech recognition for Mandarin-English CS speech. We analyze different CS specific issues such as the properties mismatches between languages in a CS language pair, the unpredictable nature of switching points, and the data scarcity problem. We exploit and improve the state-of-the-art end-to-end system by merging nonlinguistic symbols, by integrating language identification using hierarchical softmax, by modeling subword units, by artificially lowering the speaking rate, and by augmenting data using speed perturbed technique and several monolingual datasets to improve the final performance not only on CS speech but also on monolingual benchmarks in order to making the system more applicable on real life settings. Finally, we explore the effect of different language model integration methods on the performance of the proposed model. Our experimental results reveal that all the proposed techniques improve the recognition performance. The best combined system improves the baseline system by up to 35% relatively in terms of mixed error rate and delivers acceptable performance on monolingual benchmarks.
语码转换(Code-Switching, CS)是多语言社会中常见的语言现象,指的是说话时在语言之间进行转换。本文介绍了我们对汉语-英语CS语音端到端语音识别的研究。我们分析了不同的CS特定问题,如CS语言对中语言之间的属性不匹配,切换点的不可预测性以及数据稀缺性问题。我们利用和改进了最先进的端到端系统,通过合并非语言符号,通过使用分层softmax集成语言识别,通过建模子词单位,通过人为降低说话速度,通过使用速度扰动技术和几个单语数据集来增加数据,不仅在CS语音上提高最终性能,而且在单语基准上提高性能,使系统更适用于现实生活设置。最后,我们探讨了不同的语言模型集成方法对所提模型性能的影响。实验结果表明,所提方法均能提高识别性能。在混合错误率方面,最好的组合系统将基线系统提高了35%,并且在单语言基准测试中提供了可接受的性能。
{"title":"Integrating Knowledge in End-to-End Automatic Speech Recognition for Mandarin-English Code-Switching","authors":"Chia-Yu Li, Ngoc Thang Vu","doi":"10.1109/IALP48816.2019.9037688","DOIUrl":"https://doi.org/10.1109/IALP48816.2019.9037688","url":null,"abstract":"Code-Switching (CS) is a common linguistic phenomenon in multilingual communities that consists of switching between languages while speaking. This paper presents our investigations on end-to-end speech recognition for Mandarin-English CS speech. We analyze different CS specific issues such as the properties mismatches between languages in a CS language pair, the unpredictable nature of switching points, and the data scarcity problem. We exploit and improve the state-of-the-art end-to-end system by merging nonlinguistic symbols, by integrating language identification using hierarchical softmax, by modeling subword units, by artificially lowering the speaking rate, and by augmenting data using speed perturbed technique and several monolingual datasets to improve the final performance not only on CS speech but also on monolingual benchmarks in order to making the system more applicable on real life settings. Finally, we explore the effect of different language model integration methods on the performance of the proposed model. Our experimental results reveal that all the proposed techniques improve the recognition performance. The best combined system improves the baseline system by up to 35% relatively in terms of mixed error rate and delivers acceptable performance on monolingual benchmarks.","PeriodicalId":208066,"journal":{"name":"2019 International Conference on Asian Language Processing (IALP)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123306240","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Speech Command Classification System for Sinhala Language based on Automatic Speech Recognition 基于自动语音识别的僧伽罗语语音命令分类系统
Pub Date : 2019-11-01 DOI: 10.1109/IALP48816.2019.9037648
Thilini Dinushika, Lakshika Kavmini, Pamoda Abeyawardhana, Uthayasanker Thayasivam, Sanath Jayasena
Conversational Artificial Intelligence is revolutionizing the world with its power of converting the conventional computer to a human-like-computer. Exploiting the speaker’s intention is one of the major aspects in the field of conversational Artificial Intelligence. A significant challenge that hinders the effectiveness of identifying the speaker’s intention is the lack of language resources. To address this issue, we present a domain-specific speech command classification system for Sinhala, a low-resourced language. It accomplishes intent detection for the spoken Sinhala language using Automatic Speech Recognition and Natural Language Understanding. The proposed system can be effectively utilized in value-added applications such as Sinhala speech dialog systems. The system consists of an Automatic Speech Recognition engine to convert continuous natural human voice in Sinhala language to its textual representation and a text classifier to accurately understand the user intention. We also present a novel dataset for this task, 4.15 hours of Sinhala speech corpus in the banking domain. Our new Sinhala speech command classification system provides an accuracy of 89.7% in predicting the intent of an utterance. It outperforms the state-of-the-art direct speech-to-intent classification systems developed for the Sinhala language. Moreover, the Automatic Speech Recognition engine shows the Word Error Rate as 12.04% and the Sentence Error Rate as 21.56%. In addition, our experiments provide useful insights on speech-to-intent classification to researchers in low resource spoken language understanding.
对话式人工智能凭借其将传统计算机转换为类人计算机的能力正在彻底改变世界。利用说话人的意图是会话人工智能研究的主要方向之一。语言资源的缺乏是阻碍有效识别说话人意图的一个重要挑战。为了解决这个问题,我们提出了一个针对僧伽罗语的特定领域语音命令分类系统。它利用自动语音识别和自然语言理解技术实现了对僧伽罗语的意图检测。该系统可有效地用于诸如僧伽罗语语音对话系统等增值应用。该系统由自动语音识别引擎(Automatic Speech Recognition engine)和文本分类器(text classifier)组成,前者用于将连续的僧伽罗语自然人声转换为文本表示形式,后者用于准确理解用户意图。我们还为这项任务提供了一个新的数据集,即银行领域4.15小时的僧伽罗语语料库。我们的新僧伽罗语语音命令分类系统在预测话语意图方面提供了89.7%的准确率。它优于为僧伽罗语开发的最先进的直接语音到意图分类系统。此外,自动语音识别引擎显示单词错误率为12.04%,句子错误率为21.56%。此外,我们的实验为低资源口语理解的研究人员提供了语音到意图分类的有用见解。
{"title":"Speech Command Classification System for Sinhala Language based on Automatic Speech Recognition","authors":"Thilini Dinushika, Lakshika Kavmini, Pamoda Abeyawardhana, Uthayasanker Thayasivam, Sanath Jayasena","doi":"10.1109/IALP48816.2019.9037648","DOIUrl":"https://doi.org/10.1109/IALP48816.2019.9037648","url":null,"abstract":"Conversational Artificial Intelligence is revolutionizing the world with its power of converting the conventional computer to a human-like-computer. Exploiting the speaker’s intention is one of the major aspects in the field of conversational Artificial Intelligence. A significant challenge that hinders the effectiveness of identifying the speaker’s intention is the lack of language resources. To address this issue, we present a domain-specific speech command classification system for Sinhala, a low-resourced language. It accomplishes intent detection for the spoken Sinhala language using Automatic Speech Recognition and Natural Language Understanding. The proposed system can be effectively utilized in value-added applications such as Sinhala speech dialog systems. The system consists of an Automatic Speech Recognition engine to convert continuous natural human voice in Sinhala language to its textual representation and a text classifier to accurately understand the user intention. We also present a novel dataset for this task, 4.15 hours of Sinhala speech corpus in the banking domain. Our new Sinhala speech command classification system provides an accuracy of 89.7% in predicting the intent of an utterance. It outperforms the state-of-the-art direct speech-to-intent classification systems developed for the Sinhala language. Moreover, the Automatic Speech Recognition engine shows the Word Error Rate as 12.04% and the Sentence Error Rate as 21.56%. In addition, our experiments provide useful insights on speech-to-intent classification to researchers in low resource spoken language understanding.","PeriodicalId":208066,"journal":{"name":"2019 International Conference on Asian Language Processing (IALP)","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122103833","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Chinese Spelling Check based on Sequence Labeling 基于序列标注的汉语拼写检查
Pub Date : 2019-11-01 DOI: 10.1109/IALP48816.2019.9037652
Z. Han, Chengguo Lv, Qiansheng Wang, G. Fu
Chinese has been widely used by people from all over the world. Various Chinese spelling errors may occur inevitably from Chinese as Foreign Language (CFL) beginners. This paper presents a method for Chinese spelling check to detect and correct spelling errors in a sentence. Our approach is based on the sequence labeling model using the bi-direction LSTM network. We also utilize the Pinyin feature and Chinese strokes feature to improve performance. The evaluation on the SIGHAN-8 shows that our approach gets good performance on both detection and correction tasks.
汉语已被世界各地的人们广泛使用。对外汉语初学者不可避免地会出现各种汉语拼写错误。本文提出了一种汉语拼写检查方法,用于检测和纠正句子中的拼写错误。我们的方法是基于使用双向LSTM网络的序列标记模型。我们还利用拼音功能和汉字笔画功能来提高性能。对SIGHAN-8的评估表明,我们的方法在检测和校正任务上都取得了良好的性能。
{"title":"Chinese Spelling Check based on Sequence Labeling","authors":"Z. Han, Chengguo Lv, Qiansheng Wang, G. Fu","doi":"10.1109/IALP48816.2019.9037652","DOIUrl":"https://doi.org/10.1109/IALP48816.2019.9037652","url":null,"abstract":"Chinese has been widely used by people from all over the world. Various Chinese spelling errors may occur inevitably from Chinese as Foreign Language (CFL) beginners. This paper presents a method for Chinese spelling check to detect and correct spelling errors in a sentence. Our approach is based on the sequence labeling model using the bi-direction LSTM network. We also utilize the Pinyin feature and Chinese strokes feature to improve performance. The evaluation on the SIGHAN-8 shows that our approach gets good performance on both detection and correction tasks.","PeriodicalId":208066,"journal":{"name":"2019 International Conference on Asian Language Processing (IALP)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117108049","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
2019 International Conference on Asian Language Processing (IALP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1