首页 > 最新文献

International Conference on Text, Speech and Dialogue最新文献

英文 中文
Advancing Hungarian Text Processing with HuSpaCy: Efficient and Accurate NLP Pipelines 推进匈牙利语文本处理与HuSpaCy:高效和准确的NLP管道
Pub Date : 2023-08-24 DOI: 10.1007/978-3-031-40498-6_6
György Orosz, GergHo Szab'o, P'eter Berkecz, Zsolt Sz'ant'o, Richárd Farkas
{"title":"Advancing Hungarian Text Processing with HuSpaCy: Efficient and Accurate NLP Pipelines","authors":"György Orosz, GergHo Szab'o, P'eter Berkecz, Zsolt Sz'ant'o, Richárd Farkas","doi":"10.1007/978-3-031-40498-6_6","DOIUrl":"https://doi.org/10.1007/978-3-031-40498-6_6","url":null,"abstract":"","PeriodicalId":358274,"journal":{"name":"International Conference on Text, Speech and Dialogue","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129633597","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Dataset and Strong Baselines for Classification of Czech News Texts 捷克语新闻文本分类的数据集和强基线
Pub Date : 2023-07-20 DOI: 10.48550/arXiv.2307.10666
Hynek Kydl'ivcek, Jindřich Libovický
Pre-trained models for Czech Natural Language Processing are often evaluated on purely linguistic tasks (POS tagging, parsing, NER) and relatively simple classification tasks such as sentiment classification or article classification from a single news source. As an alternative, we present CZEch~NEws~Classification~dataset (CZE-NEC), one of the largest Czech classification datasets, composed of news articles from various sources spanning over twenty years, which allows a more rigorous evaluation of such models. We define four classification tasks: news source, news category, inferred author's gender, and day of the week. To verify the task difficulty, we conducted a human evaluation, which revealed that human performance lags behind strong machine-learning baselines built upon pre-trained transformer models. Furthermore, we show that language-specific pre-trained encoder analysis outperforms selected commercially available large-scale generative language models.
捷克自然语言处理的预训练模型通常在纯语言任务(词性标注、解析、NER)和相对简单的分类任务(如情感分类或来自单一新闻来源的文章分类)上进行评估。作为替代方案,我们提出了捷克~新闻~分类~数据集(CZE-NEC),这是最大的捷克分类数据集之一,由20多年来各种来源的新闻文章组成,可以对这些模型进行更严格的评估。我们定义了四个分类任务:新闻来源、新闻类别、推断作者性别和星期几。为了验证任务的难度,我们进行了人类评估,结果显示人类的表现落后于基于预训练的变压器模型建立的强大机器学习基线。此外,我们表明特定语言的预训练编码器分析优于选定的商业上可用的大规模生成语言模型。
{"title":"A Dataset and Strong Baselines for Classification of Czech News Texts","authors":"Hynek Kydl'ivcek, Jindřich Libovický","doi":"10.48550/arXiv.2307.10666","DOIUrl":"https://doi.org/10.48550/arXiv.2307.10666","url":null,"abstract":"Pre-trained models for Czech Natural Language Processing are often evaluated on purely linguistic tasks (POS tagging, parsing, NER) and relatively simple classification tasks such as sentiment classification or article classification from a single news source. As an alternative, we present CZEch~NEws~Classification~dataset (CZE-NEC), one of the largest Czech classification datasets, composed of news articles from various sources spanning over twenty years, which allows a more rigorous evaluation of such models. We define four classification tasks: news source, news category, inferred author's gender, and day of the week. To verify the task difficulty, we conducted a human evaluation, which revealed that human performance lags behind strong machine-learning baselines built upon pre-trained transformer models. Furthermore, we show that language-specific pre-trained encoder analysis outperforms selected commercially available large-scale generative language models.","PeriodicalId":358274,"journal":{"name":"International Conference on Text, Speech and Dialogue","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-07-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123936484","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Measuring Sentiment Bias in Machine Translation 机器翻译中情感偏差的测量
Pub Date : 2023-06-12 DOI: 10.48550/arXiv.2306.07152
Kai Hartung, Aaricia Herygers, Shubham Kurlekar, Khabbab Zakaria, Taylan Volkan, Sören Gröttrup, Munir Georges
Biases induced to text by generative models have become an increasingly large topic in recent years. In this paper we explore how machine translation might introduce a bias in sentiments as classified by sentiment analysis models. For this, we compare three open access machine translation models for five different languages on two parallel corpora to test if the translation process causes a shift in sentiment classes recognized in the texts. Though our statistic test indicate shifts in the label probability distributions, we find none that appears consistent enough to assume a bias induced by the translation process.
近年来,生成模型对文本产生的偏见已经成为一个越来越大的话题。在本文中,我们探讨了机器翻译如何在情感分析模型分类的情感中引入偏见。为此,我们在两个平行语料库上比较了五种不同语言的三种开放存取机器翻译模型,以测试翻译过程是否会导致文本中识别的情感类别发生变化。虽然我们的统计检验表明标签概率分布发生了变化,但我们发现没有一个数据看起来足够一致,足以假设翻译过程引起的偏差。
{"title":"Measuring Sentiment Bias in Machine Translation","authors":"Kai Hartung, Aaricia Herygers, Shubham Kurlekar, Khabbab Zakaria, Taylan Volkan, Sören Gröttrup, Munir Georges","doi":"10.48550/arXiv.2306.07152","DOIUrl":"https://doi.org/10.48550/arXiv.2306.07152","url":null,"abstract":"Biases induced to text by generative models have become an increasingly large topic in recent years. In this paper we explore how machine translation might introduce a bias in sentiments as classified by sentiment analysis models. For this, we compare three open access machine translation models for five different languages on two parallel corpora to test if the translation process causes a shift in sentiment classes recognized in the texts. Though our statistic test indicate shifts in the label probability distributions, we find none that appears consistent enough to assume a bias induced by the translation process.","PeriodicalId":358274,"journal":{"name":"International Conference on Text, Speech and Dialogue","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132962245","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Transfer Learning of Transformer-based Speech Recognition Models from Czech to Slovak 基于变压器的捷克语到斯洛伐克语语音识别模型的迁移学习
Pub Date : 2023-06-07 DOI: 10.48550/arXiv.2306.04399
Jan Lehecka, J. Psutka, J. Psutka
In this paper, we are comparing several methods of training the Slovak speech recognition models based on the Transformers architecture. Specifically, we are exploring the approach of transfer learning from the existing Czech pre-trained Wav2Vec 2.0 model into Slovak. We are demonstrating the benefits of the proposed approach on three Slovak datasets. Our Slovak models scored the best results when initializing the weights from the Czech model at the beginning of the pre-training phase. Our results show that the knowledge stored in the Cezch pre-trained model can be successfully reused to solve tasks in Slovak while outperforming even much larger public multilingual models.
在本文中,我们比较了几种基于变形金刚架构的斯洛伐克语语音识别模型的训练方法。具体来说,我们正在探索将现有的捷克语预训练的Wav2Vec 2.0模型迁移学习到斯洛伐克语的方法。我们正在三个斯洛伐克数据集上展示拟议方法的好处。在预训练阶段开始时,我们的斯洛伐克模型在初始化捷克模型的权重时获得了最好的结果。我们的结果表明,存储在捷克语预训练模型中的知识可以成功地重复使用,以解决斯洛伐克语的任务,同时优于更大的公共多语言模型。
{"title":"Transfer Learning of Transformer-based Speech Recognition Models from Czech to Slovak","authors":"Jan Lehecka, J. Psutka, J. Psutka","doi":"10.48550/arXiv.2306.04399","DOIUrl":"https://doi.org/10.48550/arXiv.2306.04399","url":null,"abstract":"In this paper, we are comparing several methods of training the Slovak speech recognition models based on the Transformers architecture. Specifically, we are exploring the approach of transfer learning from the existing Czech pre-trained Wav2Vec 2.0 model into Slovak. We are demonstrating the benefits of the proposed approach on three Slovak datasets. Our Slovak models scored the best results when initializing the weights from the Czech model at the beginning of the pre-training phase. Our results show that the knowledge stored in the Cezch pre-trained model can be successfully reused to solve tasks in Slovak while outperforming even much larger public multilingual models.","PeriodicalId":358274,"journal":{"name":"International Conference on Text, Speech and Dialogue","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125954990","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Wakeword Detection under Distribution Shifts 分布移位下的唤醒词检测
Pub Date : 2022-07-13 DOI: 10.48550/arXiv.2207.06423
S. Parthasarathi, Lu Zeng, Christin Jose, Joe Wang
We propose a novel approach for semi-supervised learning (SSL) designed to overcome distribution shifts between training and real-world data arising in the keyword spotting (KWS) task. Shifts from training data distribution are a key challenge for real-world KWS tasks: when a new model is deployed on device, the gating of the accepted data undergoes a shift in distribution, making the problem of timely updates via subsequent deployments hard. Despite the shift, we assume that the marginal distributions on labels do not change. We utilize a modified teacher/student training framework, where labeled training data is augmented with unlabeled data. Note that the teacher does not have access to the new distribution as well. To train effectively with a mix of human and teacher labeled data, we develop a teacher labeling strategy based on confidence heuristics to reduce entropy on the label distribution from the teacher model; the data is then sampled to match the marginal distribution on the labels. Large scale experimental results show that a convolutional neural network (CNN) trained on far-field audio, and evaluated on far-field audio drawn from a different distribution, obtains a 14.3% relative improvement in false discovery rate (FDR) at equal false reject rate (FRR), while yielding a 5% improvement in FDR under no distribution shift. Under a more severe distribution shift from far-field to near-field audio with a smaller fully connected network (FCN) our approach achieves a 52% relative improvement in FDR at equal FRR, while yielding a 20% relative improvement in FDR on the original distribution.
我们提出了一种新的半监督学习(SSL)方法,旨在克服关键字识别(KWS)任务中出现的训练数据和真实数据之间的分布变化。训练数据分布的变化是现实世界KWS任务的一个关键挑战:当在设备上部署新模型时,可接受数据的门接在分布上发生变化,使得通过后续部署及时更新的问题变得困难。尽管有这种变化,我们假设标签上的边际分布不变。我们使用修改后的教师/学生培训框架,其中标记的培训数据与未标记的数据相增强。请注意,老师也没有访问新分发的权限。为了有效地训练人类和教师标记数据的混合,我们开发了一种基于置信度启发式的教师标记策略,以减少来自教师模型的标签分布的熵;然后对数据进行采样以匹配标签上的边际分布。大规模实验结果表明,卷积神经网络(CNN)在远场音频上进行训练,并对来自不同分布的远场音频进行评估,在相同的错误拒绝率(FRR)下,错误发现率(FDR)相对提高14.3%,而在没有分布移动的情况下,FDR提高了5%。在从远场到近场音频的更严重的分布转移和更小的完全连接网络(FCN)下,我们的方法在相同的FRR下实现了52%的FDR相对改进,同时在原始分布上产生了20%的FDR相对改进。
{"title":"Wakeword Detection under Distribution Shifts","authors":"S. Parthasarathi, Lu Zeng, Christin Jose, Joe Wang","doi":"10.48550/arXiv.2207.06423","DOIUrl":"https://doi.org/10.48550/arXiv.2207.06423","url":null,"abstract":"We propose a novel approach for semi-supervised learning (SSL) designed to overcome distribution shifts between training and real-world data arising in the keyword spotting (KWS) task. Shifts from training data distribution are a key challenge for real-world KWS tasks: when a new model is deployed on device, the gating of the accepted data undergoes a shift in distribution, making the problem of timely updates via subsequent deployments hard. Despite the shift, we assume that the marginal distributions on labels do not change. We utilize a modified teacher/student training framework, where labeled training data is augmented with unlabeled data. Note that the teacher does not have access to the new distribution as well. To train effectively with a mix of human and teacher labeled data, we develop a teacher labeling strategy based on confidence heuristics to reduce entropy on the label distribution from the teacher model; the data is then sampled to match the marginal distribution on the labels. Large scale experimental results show that a convolutional neural network (CNN) trained on far-field audio, and evaluated on far-field audio drawn from a different distribution, obtains a 14.3% relative improvement in false discovery rate (FDR) at equal false reject rate (FRR), while yielding a 5% improvement in FDR under no distribution shift. Under a more severe distribution shift from far-field to near-field audio with a smaller fully connected network (FCN) our approach achieves a 52% relative improvement in FDR at equal FRR, while yielding a 20% relative improvement in FDR on the original distribution.","PeriodicalId":358274,"journal":{"name":"International Conference on Text, Speech and Dialogue","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125383991","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Sub 8-Bit Quantization of Streaming Keyword Spotting Models for Embedded Chipsets 嵌入式芯片组流关键字定位模型的亚8位量化
Pub Date : 2022-07-13 DOI: 10.48550/arXiv.2207.06920
Lu Zeng, S. Parthasarathi, Yuzong Liu, Alex Escott, S. Cheekatmalla, N. Strom, S. Vitaladevuni
. We propose a novel 2-stage sub 8-bit quantization aware training algorithm for all components of a 250K parameter feedforward, streaming, state-free keyword spotting model. For the 1 st -stage, we adapt a recently proposed quantization technique using a non-linear transformation with tanh ( . ) on dense layer weights. In the 2 nd -stage, we use linear quantization methods on the rest of the network, including other parameters (bias, gain, batchnorm), inputs, and activations. We conduct large scale experiments, training on 26,000 hours of de-identified production, far-field and near-field audio data (evaluating on 4,000 hours of data). We organize our results in two embedded chipset settings: a) with commodity ARM NEON instruction set and 8-bit containers, we present accuracy, CPU, and memory results using sub 8-bit weights (4, 5, 8-bit) and 8-bit quantization of rest of the network; b) with off-the-shelf neural network accelerators, for a range of weight bit widths (1 and 5-bit), while presenting accuracy results, we project reduction in memory utilization. In both configurations, our results show that the proposed algorithm can achieve: a) parity with a full floating point model’s operating point on a detection error tradeoff (DET) curve in terms of false detection rate (FDR) at false rejection rate (FRR); b) significant reduction in compute and memory, yielding up to 3 times improvement in CPU consumption and more than 4 times improvement in memory consumption.
. 针对250K参数前馈、流式、无状态关键字识别模型的所有组件,我们提出了一种新的2阶段8位次量化感知训练算法。对于第一阶段,我们采用了最近提出的使用tanh()的非线性变换的量化技术。在密集层的权重上。在第二阶段,我们对网络的其余部分使用线性量化方法,包括其他参数(偏置、增益、批范数)、输入和激活。我们进行了大规模的实验,对26000小时的去识别生产、远场和近场音频数据进行了培训(对4000小时的数据进行了评估)。我们在两种嵌入式芯片组设置中组织我们的结果:a)使用商用ARM NEON指令集和8位容器,我们使用子8位权重(4,5,8位)和网络其余部分的8位量化来呈现精度,CPU和内存结果;B)使用现成的神经网络加速器,用于权重位宽度(1位和5位)的范围,在呈现准确性结果的同时,我们预计内存利用率会降低。在两种配置下,我们的结果表明,所提出的算法可以实现:a)在误检率(FDR)和误拒率(FRR)方面,在检测误差权衡(DET)曲线上与全浮点模型的工作点的奇偶性;b)显著减少计算和内存,使CPU消耗提高3倍,内存消耗提高4倍以上。
{"title":"Sub 8-Bit Quantization of Streaming Keyword Spotting Models for Embedded Chipsets","authors":"Lu Zeng, S. Parthasarathi, Yuzong Liu, Alex Escott, S. Cheekatmalla, N. Strom, S. Vitaladevuni","doi":"10.48550/arXiv.2207.06920","DOIUrl":"https://doi.org/10.48550/arXiv.2207.06920","url":null,"abstract":". We propose a novel 2-stage sub 8-bit quantization aware training algorithm for all components of a 250K parameter feedforward, streaming, state-free keyword spotting model. For the 1 st -stage, we adapt a recently proposed quantization technique using a non-linear transformation with tanh ( . ) on dense layer weights. In the 2 nd -stage, we use linear quantization methods on the rest of the network, including other parameters (bias, gain, batchnorm), inputs, and activations. We conduct large scale experiments, training on 26,000 hours of de-identified production, far-field and near-field audio data (evaluating on 4,000 hours of data). We organize our results in two embedded chipset settings: a) with commodity ARM NEON instruction set and 8-bit containers, we present accuracy, CPU, and memory results using sub 8-bit weights (4, 5, 8-bit) and 8-bit quantization of rest of the network; b) with off-the-shelf neural network accelerators, for a range of weight bit widths (1 and 5-bit), while presenting accuracy results, we project reduction in memory utilization. In both configurations, our results show that the proposed algorithm can achieve: a) parity with a full floating point model’s operating point on a detection error tradeoff (DET) curve in terms of false detection rate (FDR) at false rejection rate (FRR); b) significant reduction in compute and memory, yielding up to 3 times improvement in CPU consumption and more than 4 times improvement in memory consumption.","PeriodicalId":358274,"journal":{"name":"International Conference on Text, Speech and Dialogue","volume":"70 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124287653","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
PoCaP Corpus: A Multimodal Dataset for Smart Operating Room Speech Assistant using Interventional Radiology Workflow Analysis PoCaP语料库:基于介入放射学工作流程分析的智能手术室语音助手多模态数据集
Pub Date : 2022-06-24 DOI: 10.48550/arXiv.2206.12320
K. Demir, M. May, A. Schmid, M. Uder, K. Breininger, T. Weise, A. Maier, Seung Hee Yang
. This paper presents a new multimodal interventional radiology dataset, called PoCaP (Port Catheter Placement) Corpus. This corpus consists of speech and audio signals in German, X-ray images, and system commands collected from 31 PoCaP interventions by six surgeons with average duration of 81 . 4 ± 41 . 0 minutes. The corpus aims to provide a resource for developing a smart speech assistant in operating rooms. In particular, it may be used to develop a speech-controlled system that enables surgeons to control the operation parameters such as C-arm movements and table positions. In order to record the dataset, we acquired consent by the institutional review board and workers’ council in the University Hospital Erlangen and by the patients for data privacy. We describe the recording set-up, data structure, workflow and preprocessing steps, and report the first PoCaP Corpus speech recognition analysis results with 11.52% word error rate using pretrained models. The findings suggest that the data has the potential to build a robust command recognition system and will allow the development of a novel intervention support systems using speech and image in the medical
。本文提出了一个新的多模式介入放射学数据集,称为PoCaP(端口导管放置)语料库。该语料库包括德语语音和音频信号、x射线图像和系统命令,这些数据来自6位外科医生平均81次的31次PoCaP干预。4±41。0分钟。该语料库旨在为手术室智能语音助手的开发提供资源。特别是,它可以用来开发一种语音控制系统,使外科医生能够控制手术参数,如c型臂运动和桌子位置。为了记录数据集,我们获得了埃尔兰根大学医院机构审查委员会和工人委员会以及患者的同意,以保护数据隐私。我们描述了记录设置、数据结构、工作流程和预处理步骤,并报告了首次使用预训练模型的PoCaP语料库语音识别分析结果,错误率为11.52%。研究结果表明,这些数据有可能建立一个强大的命令识别系统,并将允许在医疗中使用语音和图像开发一种新的干预支持系统
{"title":"PoCaP Corpus: A Multimodal Dataset for Smart Operating Room Speech Assistant using Interventional Radiology Workflow Analysis","authors":"K. Demir, M. May, A. Schmid, M. Uder, K. Breininger, T. Weise, A. Maier, Seung Hee Yang","doi":"10.48550/arXiv.2206.12320","DOIUrl":"https://doi.org/10.48550/arXiv.2206.12320","url":null,"abstract":". This paper presents a new multimodal interventional radiology dataset, called PoCaP (Port Catheter Placement) Corpus. This corpus consists of speech and audio signals in German, X-ray images, and system commands collected from 31 PoCaP interventions by six surgeons with average duration of 81 . 4 ± 41 . 0 minutes. The corpus aims to provide a resource for developing a smart speech assistant in operating rooms. In particular, it may be used to develop a speech-controlled system that enables surgeons to control the operation parameters such as C-arm movements and table positions. In order to record the dataset, we acquired consent by the institutional review board and workers’ council in the University Hospital Erlangen and by the patients for data privacy. We describe the recording set-up, data structure, workflow and preprocessing steps, and report the first PoCaP Corpus speech recognition analysis results with 11.52% word error rate using pretrained models. The findings suggest that the data has the potential to build a robust command recognition system and will allow the development of a novel intervention support systems using speech and image in the medical","PeriodicalId":358274,"journal":{"name":"International Conference on Text, Speech and Dialogue","volume":"2022 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127601420","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
TOKEN is a MASK: Few-shot Named Entity Recognition with Pre-trained Language Models TOKEN是一种掩码:使用预训练的语言模型进行少镜头命名实体识别
Pub Date : 2022-06-15 DOI: 10.48550/arXiv.2206.07841
A. Davody, David Ifeoluwa Adelani, Thomas Kleinbauer, D. Klakow
Transferring knowledge from one domain to another is of practical importance for many tasks in natural language processing, especially when the amount of available data in the target domain is limited. In this work, we propose a novel few-shot approach to domain adaptation in the context of Named Entity Recognition (NER). We propose a two-step approach consisting of a variable base module and a template module that leverages the knowledge captured in pre-trained language models with the help of simple descriptive patterns. Our approach is simple yet versatile and can be applied in few-shot and zero-shot settings. Evaluating our lightweight approach across a number of different datasets shows that it can boost the performance of state-of-the-art baselines by 2-5% F1-score.
将知识从一个领域转移到另一个领域对于自然语言处理中的许多任务具有重要的实际意义,特别是当目标领域的可用数据量有限时。在这项工作中,我们提出了一种在命名实体识别(NER)背景下进行领域自适应的新方法。我们提出了一种两步的方法,包括一个变量基础模块和一个模板模块,该模块利用在简单描述性模式的帮助下从预训练的语言模型中获取的知识。我们的方法是简单而通用的,可以应用于少拍和零拍的设置。在许多不同的数据集上评估我们的轻量级方法表明,它可以将最先进基线的性能提高2-5%的f1分数。
{"title":"TOKEN is a MASK: Few-shot Named Entity Recognition with Pre-trained Language Models","authors":"A. Davody, David Ifeoluwa Adelani, Thomas Kleinbauer, D. Klakow","doi":"10.48550/arXiv.2206.07841","DOIUrl":"https://doi.org/10.48550/arXiv.2206.07841","url":null,"abstract":"Transferring knowledge from one domain to another is of practical importance for many tasks in natural language processing, especially when the amount of available data in the target domain is limited. In this work, we propose a novel few-shot approach to domain adaptation in the context of Named Entity Recognition (NER). We propose a two-step approach consisting of a variable base module and a template module that leverages the knowledge captured in pre-trained language models with the help of simple descriptive patterns. Our approach is simple yet versatile and can be applied in few-shot and zero-shot settings. Evaluating our lightweight approach across a number of different datasets shows that it can boost the performance of state-of-the-art baselines by 2-5% F1-score.","PeriodicalId":358274,"journal":{"name":"International Conference on Text, Speech and Dialogue","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121365468","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Transformer-based Automatic Speech Recognition of Formal and Colloquial Czech in MALACH Project MALACH项目中基于变压器的正式和口语捷克语自动语音识别
Pub Date : 2022-06-15 DOI: 10.1007/978-3-031-16270-1_25
Jan Lehecka, J. Psutka, J. Psutka
{"title":"Transformer-based Automatic Speech Recognition of Formal and Colloquial Czech in MALACH Project","authors":"Jan Lehecka, J. Psutka, J. Psutka","doi":"10.1007/978-3-031-16270-1_25","DOIUrl":"https://doi.org/10.1007/978-3-031-16270-1_25","url":null,"abstract":"","PeriodicalId":358274,"journal":{"name":"International Conference on Text, Speech and Dialogue","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133790623","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Going Beyond the Cookie Theft Picture Test: Detecting Cognitive Impairments using Acoustic Features 超越饼干盗窃图片测试:使用声学特征检测认知障碍
Pub Date : 2022-06-10 DOI: 10.1007/978-3-031-16270-1_36
Franziska Braun, Andreas Erzigkeit, H. Lehfeld, T. Hillemacher, K. Riedhammer, S. Bayerl
{"title":"Going Beyond the Cookie Theft Picture Test: Detecting Cognitive Impairments using Acoustic Features","authors":"Franziska Braun, Andreas Erzigkeit, H. Lehfeld, T. Hillemacher, K. Riedhammer, S. Bayerl","doi":"10.1007/978-3-031-16270-1_36","DOIUrl":"https://doi.org/10.1007/978-3-031-16270-1_36","url":null,"abstract":"","PeriodicalId":358274,"journal":{"name":"International Conference on Text, Speech and Dialogue","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125923976","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
期刊
International Conference on Text, Speech and Dialogue
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1