首页 > 最新文献

LatinX in AI at North American Chapter of the Association for Computational Linguistics Conference 2022最新文献

英文 中文
Incorporating Natural Language Processing models in Mexico City's 311 Locatel 墨西哥城311 Locatel的自然语言处理模型
Alejandro Molina-Villegas, Edwin Aldana-Bibadilla, O. Siordia, Jorge Pérez
Natural Language Processing based technologies are transforming various sectors by facilitating new ways of providing services through Artificial Intelligence (AI). In this paper, we describe the methodology and present the challenges encountered during the creation of a Deep Learning-based model for classifying citizen service requests. Our system is able to effectively recognize among 48 categories of public services with an accuracy of 97% and was integrated into Mexico City’s 311, significantly increasing the government’s ability to provide better services.
基于自然语言处理的技术正在通过人工智能(AI)促进提供服务的新方式,从而改变各个部门。在本文中,我们描述了方法,并提出了在创建基于深度学习的公民服务请求分类模型过程中遇到的挑战。我们的系统能够有效识别48类公共服务,准确率达到97%,并被整合到墨西哥城的311中,大大提高了政府提供更好服务的能力。
{"title":"Incorporating Natural Language Processing models in Mexico City's 311 Locatel","authors":"Alejandro Molina-Villegas, Edwin Aldana-Bibadilla, O. Siordia, Jorge Pérez","doi":"10.52591/lxai202207101","DOIUrl":"https://doi.org/10.52591/lxai202207101","url":null,"abstract":"Natural Language Processing based technologies are transforming various sectors by facilitating new ways of providing services through Artificial Intelligence (AI). In this paper, we describe the methodology and present the challenges encountered during the creation of a Deep Learning-based model for classifying citizen service requests. Our system is able to effectively recognize among 48 categories of public services with an accuracy of 97% and was integrated into Mexico City’s 311, significantly increasing the government’s ability to provide better services.","PeriodicalId":350984,"journal":{"name":"LatinX in AI at North American Chapter of the Association for Computational Linguistics Conference 2022","volume":"66 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126226057","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic multi-modal processing of language and vision to assist people with visual impairments 语言和视觉的自动多模态处理,以帮助有视觉障碍的人
In recent years, the study of the intersection between vision and language modalities, specifically in visual question answering (VQA) models, has gained significant appeal due to its great potential in assistive applications for people with visual disabilities. Despite this, to date, many of the existing VQA models are nor applicable to this goal for at least three reasons. To begin with, they are designed to respond to a single question. That is, they are not able to give feedback to incomplete or incremental questions. Secondly, they only consider a single image which is neither blurred, nor poorly focused, nor poorly framed. All these problems are directly related to the loss of the visual capacity. People with visual disabilities may have trouble interacting with a visual user interface for asking questions and for taking adequate photographs. They also frequently need to read text captured by the images, and most current VQA systems fall short in this task. This work presents a PhD proposal with four lines of research that will be carried out until December 2025. It investigates techniques that increase the robustness of the VQA models. In particular we propose the integration of dialogue history, the analysis of more than one input image, and the incorporation of text recognition capabilities to the models. All of these contributions are motivated to assist people with vision problems with their day-to-day tasks.
近年来,视觉和语言模式的交叉研究,特别是视觉问答(VQA)模型的研究,因其在视觉障碍人士的辅助应用方面具有巨大的潜力而受到了极大的关注。尽管如此,到目前为止,由于至少三个原因,许多现有的VQA模型并不适用于这一目标。首先,它们被设计用来回答一个问题。也就是说,他们不能对不完整或增量的问题给出反馈。其次,他们只考虑一张既不模糊,也不对焦,也不相框的图像。所有这些问题都与视力的丧失直接相关。有视觉障碍的人可能在使用视觉用户界面提问和拍摄足够的照片时遇到困难。它们还经常需要读取图像捕获的文本,而大多数当前的VQA系统在这项任务中都做不到。这项工作提出了一个博士建议,包括四个研究方向,将在2025年12月之前进行。它研究了增加VQA模型鲁棒性的技术。我们特别提出了对话历史的集成,多幅输入图像的分析,以及将文本识别能力结合到模型中。所有这些贡献都是为了帮助有视力问题的人完成日常任务。
{"title":"Automatic multi-modal processing of language and vision to assist people with visual impairments","authors":"","doi":"10.52591/lxai202207104","DOIUrl":"https://doi.org/10.52591/lxai202207104","url":null,"abstract":"In recent years, the study of the intersection between vision and language modalities, specifically in visual question answering (VQA) models, has gained significant appeal due to its great potential in assistive applications for people with visual disabilities. Despite this, to date, many of the existing VQA models are nor applicable to this goal for at least three reasons. To begin with, they are designed to respond to a single question. That is, they are not able to give feedback to incomplete or incremental questions. Secondly, they only consider a single image which is neither blurred, nor poorly focused, nor poorly framed. All these problems are directly related to the loss of the visual capacity. People with visual disabilities may have trouble interacting with a visual user interface for asking questions and for taking adequate photographs. They also frequently need to read text captured by the images, and most current VQA systems fall short in this task. This work presents a PhD proposal with four lines of research that will be carried out until December 2025. It investigates techniques that increase the robustness of the VQA models. In particular we propose the integration of dialogue history, the analysis of more than one input image, and the incorporation of text recognition capabilities to the models. All of these contributions are motivated to assist people with vision problems with their day-to-day tasks.","PeriodicalId":350984,"journal":{"name":"LatinX in AI at North American Chapter of the Association for Computational Linguistics Conference 2022","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130157118","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Distributed Text Representations Using Transformers for Noisy Written Language 使用变压器对有噪声的书面语言进行分布式文本表示
A. Rodriguez, Pablo Rivas, G. Bejarano
This work proposes a methodology to derive latent representations for highly noisy text. Traditionally in Natural Language Processing systems, methods rely on words as the core components of a text. Unlike those, we propose a character-based approach to be robust against our target texts’ high syntactical noise. We propose pre-training a Transformer model (BERT) on different, general-purpose language tasks and using the pre-trained model to obtain a representation for an input text. Weights are transferred from one task in the pipeline to the other. Instead of tokenizing the text on a word or sub-word basis, we propose considering the text’s characters as tokens. The ultimate goal is that the representations produced prove useful for other downstream tasks on the data, such as criminal activity in marketplace platforms.
这项工作提出了一种方法来获得高噪声文本的潜在表示。在传统的自然语言处理系统中,方法依赖于单词作为文本的核心组成部分。与这些不同,我们提出了一种基于字符的方法来抵抗目标文本的高语法噪声。我们建议在不同的通用语言任务上预训练Transformer模型(BERT),并使用预训练的模型来获得输入文本的表示。权重从管道中的一个任务转移到另一个任务。我们建议将文本的字符作为标记,而不是在单词或子单词的基础上对文本进行标记。最终目标是生成的表示对数据的其他下游任务有用,例如市场平台上的犯罪活动。
{"title":"Distributed Text Representations Using Transformers for Noisy Written Language","authors":"A. Rodriguez, Pablo Rivas, G. Bejarano","doi":"10.52591/lxai202207102","DOIUrl":"https://doi.org/10.52591/lxai202207102","url":null,"abstract":"This work proposes a methodology to derive latent representations for highly noisy text. Traditionally in Natural Language Processing systems, methods rely on words as the core components of a text. Unlike those, we propose a character-based approach to be robust against our target texts’ high syntactical noise. We propose pre-training a Transformer model (BERT) on different, general-purpose language tasks and using the pre-trained model to obtain a representation for an input text. Weights are transferred from one task in the pipeline to the other. Instead of tokenizing the text on a word or sub-word basis, we propose considering the text’s characters as tokens. The ultimate goal is that the representations produced prove useful for other downstream tasks on the data, such as criminal activity in marketplace platforms.","PeriodicalId":350984,"journal":{"name":"LatinX in AI at North American Chapter of the Association for Computational Linguistics Conference 2022","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129354347","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Study of Question Answering on Legal Software Document using BERT based models 基于BERT模型的法律软件文档问答研究
Ernesto Quevedo Caballero, Mushfika Rahman, T. Cerný, Pablo Rivas, G. Bejarano
The transformer-based architectures have achieved remarkable success in several Natural Language Processing tasks, such as the Question Answering domain. Our research focuses on different transformer-based language models’ performance in software development legal domain specialized datasets for the Question Answering task. It compares the performance with the general-purpose Question Answering task. We have experimented with the PolicyQA dataset and conformed to documents regarding users’ data handling policies, which fall into the software legal domain. We used as base encoders BERT, ALBERT, RoBERTa, DistilBERT and LEGAL-BERT and compare their performance on the Question answering benchmark dataset SQuAD V2.0 and PolicyQA. Our results indicate that the performance of these models as contextual embeddings encoders in the PolicyQA dataset is significantly lower than in the SQuAD V2.0. Furthermore, we showed that surprisingly general domain BERT-based models like ALBERT and BERT obtain better performance than a more domain-specific trained model like LEGAL-BERT.
基于转换器的体系结构在一些自然语言处理任务中取得了显著的成功,例如问答领域。我们的研究重点是不同的基于转换的语言模型在软件开发法律领域的问答任务专用数据集中的性能。它将性能与通用问答任务进行比较。我们对PolicyQA数据集进行了实验,并遵循了有关用户数据处理策略的文档,这些文档属于软件法律领域。我们使用BERT、ALBERT、RoBERTa、DistilBERT和LEGAL-BERT作为基本编码器,并比较它们在问答基准数据集SQuAD V2.0和PolicyQA上的性能。我们的结果表明,这些模型在PolicyQA数据集中作为上下文嵌入编码器的性能明显低于SQuAD V2.0。此外,我们惊人地表明,一般基于BERT的领域模型,如ALBERT和BERT,比更特定于领域的训练模型,如LEGAL-BERT,获得了更好的性能。
{"title":"Study of Question Answering on Legal Software Document using BERT based models","authors":"Ernesto Quevedo Caballero, Mushfika Rahman, T. Cerný, Pablo Rivas, G. Bejarano","doi":"10.52591/lxai202207103","DOIUrl":"https://doi.org/10.52591/lxai202207103","url":null,"abstract":"The transformer-based architectures have achieved remarkable success in several Natural Language Processing tasks, such as the Question Answering domain. Our research focuses on different transformer-based language models’ performance in software development legal domain specialized datasets for the Question Answering task. It compares the performance with the general-purpose Question Answering task. We have experimented with the PolicyQA dataset and conformed to documents regarding users’ data handling policies, which fall into the software legal domain. We used as base encoders BERT, ALBERT, RoBERTa, DistilBERT and LEGAL-BERT and compare their performance on the Question answering benchmark dataset SQuAD V2.0 and PolicyQA. Our results indicate that the performance of these models as contextual embeddings encoders in the PolicyQA dataset is significantly lower than in the SQuAD V2.0. Furthermore, we showed that surprisingly general domain BERT-based models like ALBERT and BERT obtain better performance than a more domain-specific trained model like LEGAL-BERT.","PeriodicalId":350984,"journal":{"name":"LatinX in AI at North American Chapter of the Association for Computational Linguistics Conference 2022","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115545463","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Improving Language Model Fine-tuning with Information Gain Filtration 利用信息增益过滤改进语言模型微调
Javier Turek, Richard Antonello, Nicole M. Beckage, Alexander G. Huth
Language model fine-tuning is essential for modern natural language processing. The effectiveness of fine-tuning is limited by the inclusion of training examples that negatively affect performance. Here we present Information Gain Filtration, a general fine-tuning method, for improving the overall final performance of a fine-tuned model. We define Information Gain of an example as the improvement on a validation metric after training on that example. A secondary learner is then trained to approximate this quantity. During fine-tuning, this learner filters informative examples from uninformative ones. We show that our method is robust and has consistent improvement across datasets, fine-tuning tasks, and language model architectures.
语言模型的微调是现代自然语言处理的关键。微调的有效性受到包含负面影响性能的训练示例的限制。在这里,我们提出了信息增益过滤,一种通用的微调方法,用于提高微调模型的整体最终性能。我们将一个例子的信息增益定义为对该例子进行训练后验证度量的改进。然后训练一个二级学习器来接近这个数量。在微调过程中,这个学习器从无信息的例子中过滤出有信息的例子。我们证明了我们的方法是鲁棒的,并且在数据集、微调任务和语言模型架构之间具有一致的改进。
{"title":"Improving Language Model Fine-tuning with Information Gain Filtration","authors":"Javier Turek, Richard Antonello, Nicole M. Beckage, Alexander G. Huth","doi":"10.52591/lxai202207105","DOIUrl":"https://doi.org/10.52591/lxai202207105","url":null,"abstract":"Language model fine-tuning is essential for modern natural language processing. The effectiveness of fine-tuning is limited by the inclusion of training examples that negatively affect performance. Here we present Information Gain Filtration, a general fine-tuning method, for improving the overall final performance of a fine-tuned model. We define Information Gain of an example as the improvement on a validation metric after training on that example. A secondary learner is then trained to approximate this quantity. During fine-tuning, this learner filters informative examples from uninformative ones. We show that our method is robust and has consistent improvement across datasets, fine-tuning tasks, and language model architectures.","PeriodicalId":350984,"journal":{"name":"LatinX in AI at North American Chapter of the Association for Computational Linguistics Conference 2022","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130885475","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
LatinX in AI at North American Chapter of the Association for Computational Linguistics Conference 2022
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1