首页 > 最新文献

2021 16th International Joint Symposium on Artificial Intelligence and Natural Language Processing (iSAI-NLP)最新文献

英文 中文
SymSpell4Burmese: Symmetric Delete Spelling Correction Algorithm (SymSpell) for Burmese Spelling Checking SymSpell4Burmese:用于缅甸语拼写检查的对称删除拼写纠正算法(SymSpell)
Ei Phyu Phyu Mon, Ye Kyaw Thu, Than Than Yu, Aye Wai Oo
Spell checker is a crucial language tool of natural language processing (NLP) and becomes important due to the increase of text-based communication at work, information retrieval, fraud detection, search engines, social media and research areas. In this paper, automatic spelling checking for Burmese is studied by applying Symmetric Delete Spelling Correction Algorithm (SymSpell). We experimented by using an open source SymSpell python library and applied our developing Burmese spelling training corpus together with four frequency dictionaries on ten error types. For the error detection phase, the N-gram language model is used to check our developing spelling training corpus against a dictionary. For the correction phrase, SymSpell is applied to propose candidate corrections within a specified maximum edit distance from the misspelled word. After generating candidates, the best correction in the given context is automatically chosen according to the highest frequency with a minimum edit distance. We investigated the performance of each error type and studied the importance of the dictionary depending on the average term length and maximum edit distance for Burmese spell checker based on SymSpell. Moreover, we observed that syllable level segmentation with a maximum edit distance of 3 gives faster and higher quality results compared with word level segmentation results.
拼写检查器是自然语言处理(NLP)的重要语言工具,随着工作、信息检索、欺诈检测、搜索引擎、社交媒体和研究领域中基于文本的交流的增加,拼写检查器变得越来越重要。本文采用对称删除拼写校正算法(SymSpell)对缅甸语的自动拼写检查进行了研究。我们使用开源的SymSpell python库进行实验,并将我们开发的缅甸语拼写训练语料库与十个错误类型的四个频率字典一起应用。在错误检测阶段,使用N-gram语言模型对照字典检查我们正在开发的拼写训练语料库。对于更正短语,应用SymSpell在与拼写错误的单词指定的最大编辑距离内提出候选更正。在生成候选项后,根据最高频率和最小编辑距离自动选择给定上下文中的最佳校正。我们调查了每种错误类型的性能,并研究了基于SymSpell的缅甸语拼写检查器的平均词长和最大编辑距离对字典的重要性的影响。此外,我们观察到,与词级分词结果相比,最大编辑距离为3的音节级分词结果更快,质量更高。
{"title":"SymSpell4Burmese: Symmetric Delete Spelling Correction Algorithm (SymSpell) for Burmese Spelling Checking","authors":"Ei Phyu Phyu Mon, Ye Kyaw Thu, Than Than Yu, Aye Wai Oo","doi":"10.1109/iSAI-NLP54397.2021.9678171","DOIUrl":"https://doi.org/10.1109/iSAI-NLP54397.2021.9678171","url":null,"abstract":"Spell checker is a crucial language tool of natural language processing (NLP) and becomes important due to the increase of text-based communication at work, information retrieval, fraud detection, search engines, social media and research areas. In this paper, automatic spelling checking for Burmese is studied by applying Symmetric Delete Spelling Correction Algorithm (SymSpell). We experimented by using an open source SymSpell python library and applied our developing Burmese spelling training corpus together with four frequency dictionaries on ten error types. For the error detection phase, the N-gram language model is used to check our developing spelling training corpus against a dictionary. For the correction phrase, SymSpell is applied to propose candidate corrections within a specified maximum edit distance from the misspelled word. After generating candidates, the best correction in the given context is automatically chosen according to the highest frequency with a minimum edit distance. We investigated the performance of each error type and studied the importance of the dictionary depending on the average term length and maximum edit distance for Burmese spell checker based on SymSpell. Moreover, we observed that syllable level segmentation with a maximum edit distance of 3 gives faster and higher quality results compared with word level segmentation results.","PeriodicalId":339826,"journal":{"name":"2021 16th International Joint Symposium on Artificial Intelligence and Natural Language Processing (iSAI-NLP)","volume":"84 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121224281","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A Bidirectional LSTM Model for Classifying Chatbot Messages 聊天机器人消息分类的双向LSTM模型
Nunthawat Lhasiw, Nuttapong Sanglerdsinlapachai, Tanatorn Tanantong
Online channels, e.g., Facebook Messenger and Line, are widely used especially in COVID-19 pandemic. To quickly respond to their customer, chatbot system are implemented in many companies or organizations, connected to those channels. The Office of Registrar, Thammasat University also implements a chatbot to answer questions from students. An important step in the chatbot system is to know an intention of a question message. A bidirectional LSTM model is employed to classify a question message from the chatbot system into five intention classes. The experimental results shows that the obtained model yields an accuracy of 0.80 on our validation dataset.
特别是在新冠疫情期间,Facebook Messenger和Line等在线渠道被广泛使用。为了快速响应客户,许多公司或组织都实施了聊天机器人系统,连接到这些渠道。法政大学注册办公室也安装了一个聊天机器人来回答学生的问题。聊天机器人系统的一个重要步骤是了解问题信息的意图。利用双向LSTM模型将聊天机器人系统的问题信息划分为五个意向类。实验结果表明,该模型在验证数据集上的准确率为0.80。
{"title":"A Bidirectional LSTM Model for Classifying Chatbot Messages","authors":"Nunthawat Lhasiw, Nuttapong Sanglerdsinlapachai, Tanatorn Tanantong","doi":"10.1109/iSAI-NLP54397.2021.9678173","DOIUrl":"https://doi.org/10.1109/iSAI-NLP54397.2021.9678173","url":null,"abstract":"Online channels, e.g., Facebook Messenger and Line, are widely used especially in COVID-19 pandemic. To quickly respond to their customer, chatbot system are implemented in many companies or organizations, connected to those channels. The Office of Registrar, Thammasat University also implements a chatbot to answer questions from students. An important step in the chatbot system is to know an intention of a question message. A bidirectional LSTM model is employed to classify a question message from the chatbot system into five intention classes. The experimental results shows that the obtained model yields an accuracy of 0.80 on our validation dataset.","PeriodicalId":339826,"journal":{"name":"2021 16th International Joint Symposium on Artificial Intelligence and Natural Language Processing (iSAI-NLP)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123533040","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Anomaly Detection in Lexical Definitions via One-Class Classification Techniques 基于单类分类技术的词汇定义异常检测
Sawittree Jumpathong, Kanyanut Kriengket, P. Boonkwan, T. Supnithi
It takes a long time to build vocabularies and their definitions because they must be approved only by the experts in the meeting of building vocabularies and the definitions are also unstructured. To save time, we applied three techniques of classification to the experiments that are one-class SVMs, isolation forests, and local outlier factors, and also observed how well the method can suggest word definition status via the accuracy. As a result, the local outlier factors obtained the highest accuracy when they used vectors that were produced by USE. They can recognize the boundary of the approved class better and there are several approved clusters and outliers are scattered among them. Also, it is found that the detected status of definitions is both identical and opposite to the reference one. For the patterns of definition writing, the approved definitions are always written in the logical order, and start with wide or general information, then is followed by specific details, examples, and references of English terms or examples. In case of the rejected definitions, they are not always written in the logical order, and their definition patterns are also various - only Thai translation, Thai translation with related entries, parts of speech (POS), Thai translation, related entries, and English term references followed by definitions, etc.
构建词汇表及其定义需要花费很长时间,因为它们必须仅由构建词汇表会议的专家批准,而且定义也是非结构化的。为了节省时间,我们将单类支持向量机、隔离森林和局部离群因子三种分类技术应用到实验中,并观察了该方法如何通过准确率来提示单词定义状态。因此,当使用USE生成的向量时,局部离群因子获得了最高的精度。它们能较好地识别被批准类的边界,并且被批准的类有几个,离群值分散在其中。此外,我们还发现定义的检测状态与参考定义的检测状态既相同又相反。对于定义的写作模式,批准的定义总是按照逻辑顺序书写,并以广泛或一般的信息开始,然后是特定的细节、示例和对英语术语或示例的引用。对于被拒绝的定义,它们并不总是按照逻辑顺序编写,而且它们的定义模式也多种多样——只有泰语翻译、带相关条目的泰语翻译、词性(POS)、泰语翻译、相关条目和后跟定义的英语术语引用等。
{"title":"Anomaly Detection in Lexical Definitions via One-Class Classification Techniques","authors":"Sawittree Jumpathong, Kanyanut Kriengket, P. Boonkwan, T. Supnithi","doi":"10.1109/iSAI-NLP54397.2021.9678166","DOIUrl":"https://doi.org/10.1109/iSAI-NLP54397.2021.9678166","url":null,"abstract":"It takes a long time to build vocabularies and their definitions because they must be approved only by the experts in the meeting of building vocabularies and the definitions are also unstructured. To save time, we applied three techniques of classification to the experiments that are one-class SVMs, isolation forests, and local outlier factors, and also observed how well the method can suggest word definition status via the accuracy. As a result, the local outlier factors obtained the highest accuracy when they used vectors that were produced by USE. They can recognize the boundary of the approved class better and there are several approved clusters and outliers are scattered among them. Also, it is found that the detected status of definitions is both identical and opposite to the reference one. For the patterns of definition writing, the approved definitions are always written in the logical order, and start with wide or general information, then is followed by specific details, examples, and references of English terms or examples. In case of the rejected definitions, they are not always written in the logical order, and their definition patterns are also various - only Thai translation, Thai translation with related entries, parts of speech (POS), Thai translation, related entries, and English term references followed by definitions, etc.","PeriodicalId":339826,"journal":{"name":"2021 16th International Joint Symposium on Artificial Intelligence and Natural Language Processing (iSAI-NLP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129288123","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Text Summarization using Formal Argumentation 使用正式论证的文本摘要
Isada Sukprapa, Nguyen Duy Hung, T. Supnithi
Current approaches to text summarization are not genuinely interested in how competent readers perform the task often by re-constructing the arguments in the text then arriving at the summary from conclusions of acceptable arguments. This paper aims to mimic this natural path using formal argumentation techniques. Assuming the availability Argumentative Discourse Unit (ADU) graph of the given text, we build structured argumentation frameworks called S-ASPIC+ and ABA representing the text. Then we use ABA proof procedures to re-construct arguments in the text and evaluate their acceptabilities. Finally, we aggregate the conclusions of acceptable arguments. We demonstrate our approach using a dataset of argumentative micro-texts and report the results, describing comparisons to other methods.
当前的文本摘要方法并不真正关心有能力的读者如何完成任务,通常是通过重新构建文本中的论点,然后从可接受的论点的结论中得出摘要。本文旨在使用正式论证技术模拟这一自然路径。假设给定文本的可用性论证话语单元(ADU)图,我们构建了表示文本的结构化论证框架S-ASPIC+和ABA。然后,我们使用ABA证明程序对文本中的论点进行重构,并评估其可接受性。最后,我们汇总可接受的论点的结论。我们使用争论微文本数据集展示了我们的方法,并报告了结果,描述了与其他方法的比较。
{"title":"Text Summarization using Formal Argumentation","authors":"Isada Sukprapa, Nguyen Duy Hung, T. Supnithi","doi":"10.1109/iSAI-NLP54397.2021.9678183","DOIUrl":"https://doi.org/10.1109/iSAI-NLP54397.2021.9678183","url":null,"abstract":"Current approaches to text summarization are not genuinely interested in how competent readers perform the task often by re-constructing the arguments in the text then arriving at the summary from conclusions of acceptable arguments. This paper aims to mimic this natural path using formal argumentation techniques. Assuming the availability Argumentative Discourse Unit (ADU) graph of the given text, we build structured argumentation frameworks called S-ASPIC+ and ABA representing the text. Then we use ABA proof procedures to re-construct arguments in the text and evaluate their acceptabilities. Finally, we aggregate the conclusions of acceptable arguments. We demonstrate our approach using a dataset of argumentative micro-texts and report the results, describing comparisons to other methods.","PeriodicalId":339826,"journal":{"name":"2021 16th International Joint Symposium on Artificial Intelligence and Natural Language Processing (iSAI-NLP)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127740487","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Presentation Skills Training System Using Vibration Notification in a HyFlex Workshop HyFlex车间中使用振动通知的演示技巧培训系统
Mako Komatsu, Teruhiko Unoki, M. Shikida
The need for presentation skills is increasing year by year, and presentation lectures are being held in companies and universities. In particular, it is important to communicate interactively with the audience during the presentation. Currently, due to the influence of COVID-19, there are more and more opportunities for presentations in a hybrid with face-to-face and remote audiences. In a hybrid presentation, it is difficult to communicate with both audiences in the same way because there is a difference in awareness between face-to-face and remote audiences due to the influence of presence information. In this paper, we propose a method to support the awareness of remote audiences by sending vibration notifications to the presenter during the presentation in order to promote communication and support the improvement of presentation skills, and confirm the usefulness of the method.
对演讲技巧的需求正逐年增加,公司和大学都在举办演讲讲座。在演讲过程中,与听众进行互动交流尤为重要。目前,由于新冠肺炎疫情的影响,面对面和远程观众混合的演讲机会越来越多。在混合演示中,很难以相同的方式与两个受众进行沟通,因为由于在场信息的影响,面对面受众和远程受众之间的意识存在差异。在本文中,我们提出了一种方法,通过在演示过程中向演示者发送振动通知来支持远程观众的意识,以促进沟通和支持演示技巧的提高,并证实了该方法的有效性。
{"title":"Presentation Skills Training System Using Vibration Notification in a HyFlex Workshop","authors":"Mako Komatsu, Teruhiko Unoki, M. Shikida","doi":"10.1109/iSAI-NLP54397.2021.9678176","DOIUrl":"https://doi.org/10.1109/iSAI-NLP54397.2021.9678176","url":null,"abstract":"The need for presentation skills is increasing year by year, and presentation lectures are being held in companies and universities. In particular, it is important to communicate interactively with the audience during the presentation. Currently, due to the influence of COVID-19, there are more and more opportunities for presentations in a hybrid with face-to-face and remote audiences. In a hybrid presentation, it is difficult to communicate with both audiences in the same way because there is a difference in awareness between face-to-face and remote audiences due to the influence of presence information. In this paper, we propose a method to support the awareness of remote audiences by sending vibration notifications to the presenter during the presentation in order to promote communication and support the improvement of presentation skills, and confirm the usefulness of the method.","PeriodicalId":339826,"journal":{"name":"2021 16th International Joint Symposium on Artificial Intelligence and Natural Language Processing (iSAI-NLP)","volume":"89 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134109484","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Study of Levenshtein Transformer and Editor Transformer Models for Under-Resourced Languages 资源不足语言的Levenshtein变压器和Editor变压器模型研究
Transformers are the current state-of-the-art type of neural network model for dealing with sequences. Evidently, the most prominent application of these models is in text processing tasks, and the most prominent of these is machine translation. Recently, transformer-based models such as the Edit-Based Transformer with Repositioning (EDITOR) and Levenshtein Transformer (LevT) models have become popular in neural machine translation. To the best of our knowledge, there are no experiments for these two models using under-resourced languages. In this paper, we compared the performance and decoding time of the EDITOR model and the LevT model. We conducted the experiments for under-resourced language pairs, namely, Thai-to-English, Thai-to-Myanmar, English-to-Myanmar, and vice versa. The experimental results showed that the EDITOR model outperforms the LevT model in English-Thai, Thai-English and English-Myanmar language pairs whereas LevT achieves better score than EDITOR in Thai-Myanmar, Myanmar-Thai and Myanmar-English language pairs. Regarding the decoding time, EDITOR model is generally faster than the LevT model in the four language pairs. However, in the case of English-Myanmar and Myanmar-English pairs, the decoding time of EDITOR is slightly slower than the LevT model. At last, we investigated the system level performance of both models by means of compare-mt and word error rate (WER).
变压器是目前最先进的处理序列的神经网络模型。显然,这些模型最突出的应用是在文本处理任务中,其中最突出的是机器翻译。近年来,基于变压器的模型,如基于编辑的重定位变压器(EDITOR)和Levenshtein变压器(LevT)模型在神经网络机器翻译中得到了广泛的应用。据我们所知,目前还没有针对这两种模型使用资源不足语言的实验。在本文中,我们比较了EDITOR模型和LevT模型的性能和解码时间。我们对资源不足的语言对进行了实验,即泰语对英语,泰语对缅甸,英语对缅甸,反之亦然。实验结果表明,EDITOR模型在英语-泰语、泰语-英语和英语-缅甸语对上的表现优于LevT模型,而LevT模型在泰语-缅甸语、缅甸语-泰语和缅甸语-英语对上的表现优于EDITOR模型。在解码时间方面,在四种语言对中,EDITOR模型普遍比LevT模型快。然而,在英语- myanmar和缅甸- english对的情况下,EDITOR的解码时间比LevT模型稍慢。最后,我们通过比较mt和单词错误率(WER)来考察两种模型的系统级性能。
{"title":"A Study of Levenshtein Transformer and Editor Transformer Models for Under-Resourced Languages","authors":"","doi":"10.1109/iSAI-NLP54397.2021.9678159","DOIUrl":"https://doi.org/10.1109/iSAI-NLP54397.2021.9678159","url":null,"abstract":"Transformers are the current state-of-the-art type of neural network model for dealing with sequences. Evidently, the most prominent application of these models is in text processing tasks, and the most prominent of these is machine translation. Recently, transformer-based models such as the Edit-Based Transformer with Repositioning (EDITOR) and Levenshtein Transformer (LevT) models have become popular in neural machine translation. To the best of our knowledge, there are no experiments for these two models using under-resourced languages. In this paper, we compared the performance and decoding time of the EDITOR model and the LevT model. We conducted the experiments for under-resourced language pairs, namely, Thai-to-English, Thai-to-Myanmar, English-to-Myanmar, and vice versa. The experimental results showed that the EDITOR model outperforms the LevT model in English-Thai, Thai-English and English-Myanmar language pairs whereas LevT achieves better score than EDITOR in Thai-Myanmar, Myanmar-Thai and Myanmar-English language pairs. Regarding the decoding time, EDITOR model is generally faster than the LevT model in the four language pairs. However, in the case of English-Myanmar and Myanmar-English pairs, the decoding time of EDITOR is slightly slower than the LevT model. At last, we investigated the system level performance of both models by means of compare-mt and word error rate (WER).","PeriodicalId":339826,"journal":{"name":"2021 16th International Joint Symposium on Artificial Intelligence and Natural Language Processing (iSAI-NLP)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133951787","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Clutter Removal Algorithm Based on Grid Density with a Recursive Approach for Rockfall Detection in 3D point clouds from a Terrestrial LiDAR Scanner 基于网格密度递归的地面激光雷达三维点云落石检测杂波去除算法
Phuriphan Prathipasen, Pitisit Dillon, P. Aimmanee, Suree Teerarungsigul, Sasiwimol Nawawitphisit, S. Keerativittayanun, Jessada Karnjana
In analyzing 3D point clouds obtained from a terrestrial LiDAR scanner for rockfall detection, a widely-used clutter removal algorithm is Nearest Neighbor Clutter Removal (NNCR). However, there is a critical problem regarding computational complexity of NNCR. Subsequently, we presented a new algorithm for clutter removal based on grid density as a solution to this problem. Nevertheless, the previously proposed method showed that data points were lost. This study proposes a multi-scale grid-density-based method, assuming that the clutter is normally distributed. Outcomes from the experiment indicate that a proposed method could retrieve data points lost in the previous method. The balanced accuracies, recalls, and F-scores of the proposed method were improved by approximately 13, 33, and 17 percent, respectively, compared with the previously proposed method. Also, the proposed method is about 19 times faster than NNCR.
在对地面激光雷达扫描仪获得的三维点云进行岩落检测时,最近邻杂波去除(NNCR)是一种广泛使用的杂波去除算法。然而,关于NNCR的计算复杂性存在一个关键问题。针对这一问题,本文提出了一种基于网格密度的杂波去除算法。然而,先前提出的方法显示数据点丢失。本研究提出了一种基于多尺度网格密度的方法,假设杂波为正态分布。实验结果表明,该方法可以有效地恢复原方法中丢失的数据点。与之前提出的方法相比,该方法的平衡准确性、召回率和f分数分别提高了约13%、33%和17%。此外,该方法比NNCR快19倍左右。
{"title":"Clutter Removal Algorithm Based on Grid Density with a Recursive Approach for Rockfall Detection in 3D point clouds from a Terrestrial LiDAR Scanner","authors":"Phuriphan Prathipasen, Pitisit Dillon, P. Aimmanee, Suree Teerarungsigul, Sasiwimol Nawawitphisit, S. Keerativittayanun, Jessada Karnjana","doi":"10.1109/iSAI-NLP54397.2021.9678185","DOIUrl":"https://doi.org/10.1109/iSAI-NLP54397.2021.9678185","url":null,"abstract":"In analyzing 3D point clouds obtained from a terrestrial LiDAR scanner for rockfall detection, a widely-used clutter removal algorithm is Nearest Neighbor Clutter Removal (NNCR). However, there is a critical problem regarding computational complexity of NNCR. Subsequently, we presented a new algorithm for clutter removal based on grid density as a solution to this problem. Nevertheless, the previously proposed method showed that data points were lost. This study proposes a multi-scale grid-density-based method, assuming that the clutter is normally distributed. Outcomes from the experiment indicate that a proposed method could retrieve data points lost in the previous method. The balanced accuracies, recalls, and F-scores of the proposed method were improved by approximately 13, 33, and 17 percent, respectively, compared with the previously proposed method. Also, the proposed method is about 19 times faster than NNCR.","PeriodicalId":339826,"journal":{"name":"2021 16th International Joint Symposium on Artificial Intelligence and Natural Language Processing (iSAI-NLP)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133625268","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Coreference Resolution and Meaning Representation in a Legislative Corpus 立法语料库中的共涉决议与意义表示
Surawat Pothong, N. Facundes
This paper addresses the application and integration of coreferences resolution tasks in a legislative corpus by using SpanBERT, which is an improvement of the BERT (Bidirectional Encoder Representations from Transformers) model and semantic extraction by Abstract Meaning Representation (AMR) for reducing text complexity, meaning preservation and further applications. Our main processes are divided into four subparts: legal text pre-processing, coreference resolution, AMR, evaluation for meaning preservation, and complexity reduction. Smatch evaluation tool and Bilingual Evaluation Understudy (BLEU) scores are applied to evaluate overlapped meaning between resolved and unresolved coreference sentences. The AMR graphs after complexity have been reduced can be applied for further processing tasks with Neural Network such as legal inferencing and legal engineering tasks.
本文讨论了在立法语料库中使用SpanBERT(双向编码器表示)模型和抽象意义表示(AMR)的语义提取的改进,以降低文本复杂性、意义保留和进一步的应用)来解决共同引用解析任务的应用和集成。我们的主要过程分为四个子部分:法律文本预处理、共同参考解析、AMR、意义保留评估和复杂性降低。使用Smatch评价工具和双语评价替代(BLEU)分数来评价已解决和未解决的共指句之间的重叠意义。降低复杂度后的AMR图可以应用于神经网络的进一步处理任务,如法律推理和法律工程任务。
{"title":"Coreference Resolution and Meaning Representation in a Legislative Corpus","authors":"Surawat Pothong, N. Facundes","doi":"10.1109/iSAI-NLP54397.2021.9678168","DOIUrl":"https://doi.org/10.1109/iSAI-NLP54397.2021.9678168","url":null,"abstract":"This paper addresses the application and integration of coreferences resolution tasks in a legislative corpus by using SpanBERT, which is an improvement of the BERT (Bidirectional Encoder Representations from Transformers) model and semantic extraction by Abstract Meaning Representation (AMR) for reducing text complexity, meaning preservation and further applications. Our main processes are divided into four subparts: legal text pre-processing, coreference resolution, AMR, evaluation for meaning preservation, and complexity reduction. Smatch evaluation tool and Bilingual Evaluation Understudy (BLEU) scores are applied to evaluate overlapped meaning between resolved and unresolved coreference sentences. The AMR graphs after complexity have been reduced can be applied for further processing tasks with Neural Network such as legal inferencing and legal engineering tasks.","PeriodicalId":339826,"journal":{"name":"2021 16th International Joint Symposium on Artificial Intelligence and Natural Language Processing (iSAI-NLP)","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130233770","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Supervised Bilingual Word Embeddings for Low-Resource Language Pairs: Myanmar and Thai 低资源语言对的有监督双语词嵌入:缅甸语和泰语
Bilingual word embeddings (BWEs) represent the lexicons of two different languages in a shared embedding space, which are useful for cross-lingual natural language processing (NLP) tasks. In particular, bilingual word embeddings are extremely useful for machine translation of low-resource languages due to the rare availability of parallel corpus for that languages. Most of the researchers have already learned bilingual word embeddings for high-resource language pairs. To the best of our knowledge, there are no studies on bilingual word embeddings for low resource language pairs, Myanmar-Thai and Myanmar-English. In this paper, we present and evaluate the bilingual word embeddings for Myanmar-Thai, Myanmar-English, Thai-English, and English-Thai language pairs. To train bilingual word embeddings for each language pair, firstly, we used monolingual corpora for constructing monolingual word embeddings. A bilingual dictionary was also utilized to alleviate the problem of learning bilingual mappings as a supervised machine learning task, where a vector space is first learned independently on a monolingual corpus. Then, a linear alignment strategy is used to map the monolingual embeddings to a common bilingual vector space. Either word2vec or fastText model was used to construct monolingual word embeddings. We used bilingual dictionary induction as the intrinsic testbed for evaluating the quality of cross-lingual mappings from our constructed bilingual word embeddings. For all low-resource language pairs, monolingual word2vec embedding models with the CSLS metric achieved the best coverage and accuracy.
双语词嵌入(BWEs)表示共享嵌入空间中两种不同语言的词库,对跨语言自然语言处理(NLP)任务非常有用。特别是,双语词嵌入对于低资源语言的机器翻译非常有用,因为这种语言的平行语料很少。大多数研究人员已经学习了高资源语言对的双语词嵌入。据我们所知,目前还没有针对低资源语言对(缅泰和缅英)的双语词嵌入的研究。在本文中,我们介绍并评估了缅泰、缅英、泰英和英泰语言对的双语词嵌入。为了训练每对语言的双语词嵌入,我们首先使用单语语料库来构建单语词嵌入。我们还使用了双语词典,以减轻作为监督机器学习任务的双语映射学习的问题,即首先在单语语料库上独立学习一个向量空间。然后,使用线性对齐策略将单语嵌入映射到通用的双语向量空间。我们使用 word2vec 或 fastText 模型来构建单语词嵌入。我们使用双语词典归纳法作为内在测试平台,以评估根据我们构建的双语词嵌入进行跨语言映射的质量。对于所有低资源语言对,采用 CSLS 指标的单语 word2vec 嵌入模型实现了最佳的覆盖率和准确率。
{"title":"Supervised Bilingual Word Embeddings for Low-Resource Language Pairs: Myanmar and Thai","authors":"","doi":"10.1109/iSAI-NLP54397.2021.9678157","DOIUrl":"https://doi.org/10.1109/iSAI-NLP54397.2021.9678157","url":null,"abstract":"Bilingual word embeddings (BWEs) represent the lexicons of two different languages in a shared embedding space, which are useful for cross-lingual natural language processing (NLP) tasks. In particular, bilingual word embeddings are extremely useful for machine translation of low-resource languages due to the rare availability of parallel corpus for that languages. Most of the researchers have already learned bilingual word embeddings for high-resource language pairs. To the best of our knowledge, there are no studies on bilingual word embeddings for low resource language pairs, Myanmar-Thai and Myanmar-English. In this paper, we present and evaluate the bilingual word embeddings for Myanmar-Thai, Myanmar-English, Thai-English, and English-Thai language pairs. To train bilingual word embeddings for each language pair, firstly, we used monolingual corpora for constructing monolingual word embeddings. A bilingual dictionary was also utilized to alleviate the problem of learning bilingual mappings as a supervised machine learning task, where a vector space is first learned independently on a monolingual corpus. Then, a linear alignment strategy is used to map the monolingual embeddings to a common bilingual vector space. Either word2vec or fastText model was used to construct monolingual word embeddings. We used bilingual dictionary induction as the intrinsic testbed for evaluating the quality of cross-lingual mappings from our constructed bilingual word embeddings. For all low-resource language pairs, monolingual word2vec embedding models with the CSLS metric achieved the best coverage and accuracy.","PeriodicalId":339826,"journal":{"name":"2021 16th International Joint Symposium on Artificial Intelligence and Natural Language Processing (iSAI-NLP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131338715","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Speaker Recognition by Combining Features for Myanmar Weather Forecast Domain 结合缅甸天气预报域特征的说话人识别
Khaing Zar Mon, Reenu, Ye Kyaw Thu
Nowadays, speaker recognition has become one of the important application area of digital signal processing. Speech corpus is important in developing the speech processing and the development of the corpus is essential for low-resourced languages. Burmese (Myanmar language) can be recognized as a low-resourced language because of lack of available resources for speech processing research. Speaker Recognition (SReg) is an approach used to automatically recognize a speaker from their speech utterance. The main concept of SReg is to recognize the information of the speaker’s identity. In SReg, various features have been extracted to reflect the characteristics of the speakers. In this paper, an effective multi-feature combination and comparison of performance between the different size of data are proposed. In this work, weather news from Department of Meteorology and Hydrology, Myanmar is collected. The total size of the implemented Burmese speech corpus is over 10 hours and it contained 13 females and 3 males. The dataset is split into training data and testing data in 4:1 ratio. The experimental results on 16 speakers show that the proposed Burmese speaker recognition based on multi-feature combination achieved 99.16% accuracy and high applicability.
现如今,说话人识别已经成为数字信号处理的重要应用领域之一。语音语料库对语音处理的发展具有重要意义,对于资源匮乏的语言,语料库的开发是必不可少的。由于缺乏可用于语音处理研究的资源,缅甸语可以被认为是一种低资源语言。说话人识别(SReg)是一种从说话人的话语中自动识别说话人的方法。SReg的主要概念是识别说话人的身份信息。在SReg中,提取了各种特征来反映说话人的特征。本文提出了一种有效的多特征组合方法,并对不同大小的数据进行了性能比较。在这项工作中,收集了缅甸气象和水文部门的天气新闻。所实施的缅甸语语料库总规模超过10小时,其中包含13名女性和3名男性。数据集按4:1的比例分为训练数据和测试数据。对16名说话人的实验结果表明,基于多特征组合的缅甸语说话人识别准确率达到99.16%,具有较高的适用性。
{"title":"Speaker Recognition by Combining Features for Myanmar Weather Forecast Domain","authors":"Khaing Zar Mon, Reenu, Ye Kyaw Thu","doi":"10.1109/iSAI-NLP54397.2021.9678153","DOIUrl":"https://doi.org/10.1109/iSAI-NLP54397.2021.9678153","url":null,"abstract":"Nowadays, speaker recognition has become one of the important application area of digital signal processing. Speech corpus is important in developing the speech processing and the development of the corpus is essential for low-resourced languages. Burmese (Myanmar language) can be recognized as a low-resourced language because of lack of available resources for speech processing research. Speaker Recognition (SReg) is an approach used to automatically recognize a speaker from their speech utterance. The main concept of SReg is to recognize the information of the speaker’s identity. In SReg, various features have been extracted to reflect the characteristics of the speakers. In this paper, an effective multi-feature combination and comparison of performance between the different size of data are proposed. In this work, weather news from Department of Meteorology and Hydrology, Myanmar is collected. The total size of the implemented Burmese speech corpus is over 10 hours and it contained 13 females and 3 males. The dataset is split into training data and testing data in 4:1 ratio. The experimental results on 16 speakers show that the proposed Burmese speaker recognition based on multi-feature combination achieved 99.16% accuracy and high applicability.","PeriodicalId":339826,"journal":{"name":"2021 16th International Joint Symposium on Artificial Intelligence and Natural Language Processing (iSAI-NLP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132557019","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2021 16th International Joint Symposium on Artificial Intelligence and Natural Language Processing (iSAI-NLP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1