首页 > 最新文献

Computer Speech and Language最新文献

英文 中文
Joint speaker encoder and neural back-end model for fully end-to-end automatic speaker verification with multiple enrollment utterances 联合扬声器编码器和神经后端模型,实现具有多个登记语料的完全端到端自动扬声器验证
IF 4.3 3区 计算机科学 Q1 Mathematics Pub Date : 2024-01-18 DOI: 10.1016/j.csl.2024.101619
Chang Zeng , Xiaoxiao Miao , Xin Wang , Erica Cooper , Junichi Yamagishi

Conventional automatic speaker verification systems can usually be decomposed into a front-end model such as time delay neural network (TDNN) for extracting speaker embeddings and a back-end model such as statistics-based probabilistic linear discriminant analysis (PLDA) or neural network-based neural PLDA (NPLDA) for similarity scoring. However, the sequential optimization of the front-end and back-end models may lead to a local minimum, which theoretically prevents the whole system from achieving the best optimization. Although some methods have been proposed for jointly optimizing the two models, such as the generalized end-to-end (GE2E) model and NPLDA E2E model, most of these methods have not fully investigated how to model the intra-relationship between multiple enrollment utterances. In this paper, we propose a new E2E joint method for speaker verification especially designed for the practical scenario of multiple enrollment utterances. To leverage the intra-relationship among multiple enrollment utterances, our model comes equipped with frame-level and utterance-level attention mechanisms. Additionally, focal loss is utilized to balance the importance of positive and negative samples within a mini-batch and focus on the difficult samples during the training process. We also utilize several data augmentation techniques, including conventional noise augmentation using MUSAN and RIRs datasets and a unique speaker embedding-level mixup strategy for better optimization.

传统的自动语音验证系统通常可分解为用于提取说话人嵌入的前端模型(如时延神经网络(TDNN))和用于相似性评分的后端模型(如基于统计的概率线性判别分析(PLDA)或基于神经网络的神经线性判别分析(NPLDA))。然而,前端和后端模型的顺序优化可能会导致局部最小值,这在理论上会阻碍整个系统实现最佳优化。虽然已经提出了一些对两个模型进行联合优化的方法,如广义端到端(GE2E)模型和 NPLDA E2E 模型,但这些方法大多没有充分研究如何对多个报名语篇之间的内在关系进行建模。在本文中,我们提出了一种新的 E2E 联合方法来验证说话人,这种方法是专门针对多报名语料的实际场景而设计的。为了充分利用多个注册语篇之间的内在关系,我们的模型配备了帧级和语篇级关注机制。此外,在训练过程中,我们还利用焦点损失(focal loss)来平衡迷你批次中正负样本的重要性,并将重点放在困难样本上。我们还采用了多种数据增强技术,包括使用 MUSAN 和 RIRs 数据集的传统噪声增强技术,以及独特的扬声器嵌入级混合策略,以实现更好的优化。
{"title":"Joint speaker encoder and neural back-end model for fully end-to-end automatic speaker verification with multiple enrollment utterances","authors":"Chang Zeng ,&nbsp;Xiaoxiao Miao ,&nbsp;Xin Wang ,&nbsp;Erica Cooper ,&nbsp;Junichi Yamagishi","doi":"10.1016/j.csl.2024.101619","DOIUrl":"10.1016/j.csl.2024.101619","url":null,"abstract":"<div><p>Conventional automatic speaker verification systems can usually be decomposed into a front-end model such as time delay neural network (TDNN) for extracting speaker embeddings and a back-end model such as statistics-based probabilistic linear discriminant analysis (PLDA) or neural network-based neural PLDA (NPLDA) for similarity scoring. However, the sequential optimization of the front-end and back-end models may lead to a local minimum, which theoretically prevents the whole system from achieving the best optimization. Although some methods have been proposed for jointly optimizing the two models, such as the generalized end-to-end (GE2E) model and NPLDA E2E model, most of these methods have not fully investigated how to model the intra-relationship between multiple enrollment utterances. In this paper, we propose a new E2E joint method for speaker verification especially designed for the practical scenario of multiple enrollment utterances. To leverage the intra-relationship among multiple enrollment utterances, our model comes equipped with frame-level and utterance-level attention mechanisms. Additionally, focal loss is utilized to balance the importance of positive and negative samples within a mini-batch and focus on the difficult samples during the training process. We also utilize several data augmentation techniques, including conventional noise augmentation using MUSAN and RIRs datasets and a unique speaker embedding-level mixup strategy for better optimization.</p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-01-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0885230824000020/pdfft?md5=ef4d8f62c6e421e3a3accd1ee4ea9a64&pid=1-s2.0-S0885230824000020-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139507969","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Scale-aware dual-branch complex convolutional recurrent network for monaural speech enhancement 用于单声道语音增强的规模感知双分支复杂卷积递归网络
IF 4.3 3区 计算机科学 Q1 Mathematics Pub Date : 2024-01-13 DOI: 10.1016/j.csl.2024.101618
Yihao Li , Meng Sun , Xiongwei Zhang , Hugo Van hamme

A key step to single channel speech enhancement is the orthogonal separation of speech and noise. In this paper, a dual branch complex convolutional recurrent network (DBCCRN) is proposed to separate the complex spectrograms of speech and noises simultaneously. To model both local and global information, we incorporate conformer modules into our network. The orthogonality of the outputs of the two branches can be improved by optimizing the Signal-to-Noise Ratio (SNR) related losses. However, we found the models trained by two existing versions of SI-SNRs will yield enhanced speech at a very different scale from that of its clean counterpart. SNR loss will lead to a shrink amplitude of enhanced speech as well. A solution to this problem is to simply normalize the output, but it only works for off-line processing, not for the streaming one. When streaming speech enhancement is required, the error scale will lead to the degradation of speech quality. From an analytical inspection of the weakness of the models trained by SNR and SI-SNR losses, a new loss function called scale-aware SNR (SA-SNR) is proposed to cope with the scale variations of the enhanced speech. SA-SNR improves over SI-SNR by introducing an extra regularization term that encourages the model to produce signals of similar scale as the input, which has little influence on the perceptual quality of the enhanced speech. In addition, the commonly used evaluation recipe for speech enhancement may not be sufficient to comprehensively reflect the performance of the speech enhancement methods using SI-SNR losses, where amplitude variations of input speech should be carefully considered. A new evaluation recipe called ScaleError is introduced. Experiments show that our proposed method improves over the existing baselines on the evaluation sets of the voice bank corpus, DEMAND and the Interspeech 2020 Deep Noise Suppression Challenge, by obtaining higher scores for PESQ, STOI, SSNR, CSIG, CBAK and COVL.

单通道语音增强的关键步骤是语音和噪声的正交分离。本文提出了一种双分支复杂卷积递归网络(DBCCRN),可同时分离语音和噪声的复杂频谱图。为了对局部和全局信息进行建模,我们在网络中加入了共形模块。通过优化信噪比(SNR)相关损失,可以改善两个分支输出的正交性。然而,我们发现,由现有的两个版本的 SI-SNR 训练出的模型所产生的增强语音,其规模与纯净语音的规模截然不同。信噪比损失也会导致增强语音的振幅缩小。解决这一问题的办法是简单地对输出进行归一化处理,但这只适用于离线处理,不适用于流式处理。当需要进行流式语音增强时,误差标度将导致语音质量下降。通过对用 SNR 和 SI-SNR 损失训练的模型的弱点进行分析,提出了一种新的损失函数,称为尺度感知 SNR(SA-SNR),以应对增强语音的尺度变化。与 SI-SNR 相比,SA-SNR 引入了一个额外的正则化项,鼓励模型产生与输入信号相似的信号,这对增强语音的感知质量影响很小。此外,常用的语音增强评估方法可能不足以全面反映使用 SI-SNR 损失的语音增强方法的性能,在这种情况下,应仔细考虑输入语音的振幅变化。我们引入了一种名为 ScaleError 的新评估方法。实验表明,在语音库语料、DEMAND 和 Interspeech 2020 深度噪声抑制挑战赛的评估集上,我们提出的方法在 PESQ、STOI、SSNR、CSIG、CBAK 和 COVL 方面获得了更高的分数,比现有的基线方法有所改进。
{"title":"Scale-aware dual-branch complex convolutional recurrent network for monaural speech enhancement","authors":"Yihao Li ,&nbsp;Meng Sun ,&nbsp;Xiongwei Zhang ,&nbsp;Hugo Van hamme","doi":"10.1016/j.csl.2024.101618","DOIUrl":"https://doi.org/10.1016/j.csl.2024.101618","url":null,"abstract":"<div><p><span><span><span>A key step to single channel speech enhancement is the orthogonal separation of speech and noise. In this paper, a dual branch complex convolutional recurrent network<span> (DBCCRN) is proposed to separate the complex spectrograms of speech and noises simultaneously. To model both local and global information, we incorporate </span></span>conformer<span><span> modules into our network. The orthogonality of the outputs of the two branches can be improved by optimizing the Signal-to-Noise Ratio (SNR) related losses. However, we found the models trained by two existing versions of SI-SNRs will yield enhanced speech at a very different scale from that of its clean counterpart. SNR loss will lead to a shrink amplitude of enhanced speech as well. A solution to this problem is to simply normalize the output, but it only works for off-line processing, not for the streaming one. When streaming speech enhancement is required, the error scale will lead to the degradation of speech quality. From an analytical inspection of the weakness of the models trained by SNR and SI-SNR losses, a new loss function called scale-aware SNR (SA-SNR) is proposed to cope with the scale variations of the enhanced speech. SA-SNR improves over SI-SNR by introducing an extra </span>regularization term that encourages the model to produce signals of similar scale as the input, which has little influence on the </span></span>perceptual quality of the enhanced speech. In addition, the commonly used evaluation recipe for speech enhancement may not be sufficient to comprehensively reflect the performance of the speech enhancement methods using SI-SNR losses, where amplitude variations of input speech should be carefully considered. A new evaluation recipe called </span><em>ScaleError</em> is introduced. Experiments show that our proposed method improves over the existing baselines on the evaluation sets of the <em>voice bank corpus, DEMAND</em> and <span><em>the Interspeech 2020 Deep </em><em>Noise Suppression</em><em> Challenge</em></span>, by obtaining higher scores for PESQ, STOI, SSNR, CSIG, CBAK and COVL.</p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-01-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139487878","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A tag-based methodology for the detection of user repair strategies in task-oriented conversational agents 基于标签的任务导向型对话代理用户修复策略检测方法
IF 4.3 3区 计算机科学 Q1 Mathematics Pub Date : 2024-01-08 DOI: 10.1016/j.csl.2023.101603
Francesca Alloatti , Francesca Grasso , Roger Ferrod , Giovanni Siragusa , Luigi Di Caro , Federica Cena

Mutual comprehension is a crucial component that makes a conversation succeed. While it can be easily reached through the cooperation of the parties in human–human dialogues, such cooperation is often lacking in human–computer interaction due to technical problems, leading to broken conversations. Our goal is to work towards an effective detection of breakdowns in a conversation between humans and Conversational Agents (CA), as well as the different repair strategies users adopt when such communication problems occur. In this work, we propose a novel tag system designed to map and classify users’ repair attempts while interacting with a CA. We subsequently present a set of Machine Learning models1 trained to automatize the detection of such repair strategies. The tags are employed in a manual annotation exercise, performed on a publicly available dataset 2 of text-based task-oriented conversations. The batch of annotated data was then used to train the neural network-based classifiers. The analysis of the annotations provides interesting insights about users’ behaviour when dealing with breakdowns in a task-oriented dialogue system. The encouraging results obtained from neural models confirm the possibility of automatically recognizing occurrences of misunderstanding between users and CAs on the fly.

相互理解是对话成功的关键因素。在人与人的对话中,通过双方的合作可以很容易地实现相互理解,但在人机交互中,由于技术问题,这种合作往往会缺失,从而导致对话中断。我们的目标是致力于有效检测人类与对话代理(CA)之间对话的中断情况,以及用户在出现此类交流问题时所采取的不同修复策略。在这项工作中,我们提出了一个新颖的标签系统,旨在对用户与 CA 交互时的修复尝试进行映射和分类。随后,我们提出了一套经过训练的机器学习模型1 ,用于自动检测此类修复策略。这些标签被用于人工标注工作,该工作是在一个公开可用的基于任务导向的文本会话数据集 2 上进行的。批量注释数据随后被用于训练基于神经网络的分类器。通过对注释的分析,我们可以深入了解用户在任务导向对话系统中处理故障时的行为。神经网络模型获得的令人鼓舞的结果证实了自动识别用户与 CA 之间误解的可能性。
{"title":"A tag-based methodology for the detection of user repair strategies in task-oriented conversational agents","authors":"Francesca Alloatti ,&nbsp;Francesca Grasso ,&nbsp;Roger Ferrod ,&nbsp;Giovanni Siragusa ,&nbsp;Luigi Di Caro ,&nbsp;Federica Cena","doi":"10.1016/j.csl.2023.101603","DOIUrl":"https://doi.org/10.1016/j.csl.2023.101603","url":null,"abstract":"<div><p><span><span>Mutual comprehension is a crucial component that makes a conversation succeed. While it can be easily reached through the cooperation of the parties in human–human dialogues, such cooperation is often lacking in human–computer interaction due to technical problems, leading to broken conversations. Our goal is to work towards an effective detection of breakdowns in a conversation between humans and Conversational Agents (CA), as well as the different repair strategies users adopt when such communication problems occur. In this work, we propose a novel tag system designed to map and classify users’ repair attempts while interacting with a CA. We subsequently present a set of </span>Machine Learning models</span><span><sup>1</sup></span> trained to automatize the detection of such repair strategies. The tags are employed in a manual annotation exercise, performed on a publicly available dataset <span><sup>2</sup></span><span> of text-based task-oriented conversations. The batch of annotated data was then used to train the neural network-based classifiers. The analysis of the annotations provides interesting insights about users’ behaviour when dealing with breakdowns in a task-oriented dialogue system<span>. The encouraging results obtained from neural models confirm the possibility of automatically recognizing occurrences of misunderstanding between users and CAs on the fly.</span></span></p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-01-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139406502","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TTK: A toolkit for Tunisian linguistic analysis TTK:突尼斯语言分析工具包
IF 4.3 3区 计算机科学 Q1 Mathematics Pub Date : 2024-01-03 DOI: 10.1016/j.csl.2023.101617
Asma Mekki, Inès Zribi, Mariem Ellouze, Lamia Hadrich Belguith

Over the last two decades, many efforts have been made to provide resources to support the Arabic Natural Language Processing (NLP). Some of these resources target specific NLP tasks such as word tokenization, parsing, or sentiment analysis, while others attempt to tackle numerous tasks at once. In this paper, we present ¡¡TTK¿¿, a toolkit for Tunisian linguistic analysis. It consists of a collection of linguistic analysis tools for orthographic normalization, sentence boundaries detection, word tokenization, morphological analysis, parsing and named entity recognition. This paper focuses on the design and implementation of TTK tools.

在过去的二十年中,人们为提供阿拉伯语自然语言处理(NLP)资源做出了许多努力。其中一些资源针对特定的 NLP 任务,如单词标记化、解析或情感分析,而另一些资源则试图同时处理多项任务。在本文中,我们介绍了突尼斯语言分析工具包 "TTK"。它由一系列语言分析工具组成,包括正字法规范化、句子边界检测、单词标记化、形态分析、解析和命名实体识别。本文重点介绍 TTK 工具的设计和实施。
{"title":"TTK: A toolkit for Tunisian linguistic analysis","authors":"Asma Mekki,&nbsp;Inès Zribi,&nbsp;Mariem Ellouze,&nbsp;Lamia Hadrich Belguith","doi":"10.1016/j.csl.2023.101617","DOIUrl":"10.1016/j.csl.2023.101617","url":null,"abstract":"<div><p><span><span>Over the last two decades, many efforts have been made to provide resources to support the Arabic Natural Language Processing (NLP). Some of these resources target specific NLP tasks such as word tokenization, </span>parsing, or </span>sentiment analysis<span><span>, while others attempt to tackle numerous tasks at once. In this paper, we present ¡¡TTK¿¿, a toolkit for Tunisian linguistic analysis. It consists of a collection of linguistic analysis tools for orthographic normalization, sentence boundaries detection, word tokenization, morphological analysis, parsing and </span>named entity recognition. This paper focuses on the design and implementation of TTK tools.</span></p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139094553","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhanced local knowledge with proximity values and syntax-clusters for aspect-level sentiment analysis 利用近似值和语法簇增强本地知识,进行方面级情感分析
IF 4.3 3区 计算机科学 Q1 Mathematics Pub Date : 2023-12-28 DOI: 10.1016/j.csl.2023.101616
Pengfei Chen , Biqing Zeng , Yuwu Lu , Yun Xue , Fei Fan , Mayi Xu , Lingcong Feng

Aspect-level sentiment analysis (ALSA) aims to extract the polarity of different aspect terms in a sentence. Previous works leveraging traditional dependency syntax parsing trees (DSPT) to encode contextual syntactic information had obtained state-of-the-art results. However, these works may not be able to learn fine-grained syntactic knowledge efficiently, which makes them difficult to take advantage of local context. Furthermore, these works failed to exploit the dependency relation from DSPT sufficiently. To solve these problems, we propose a novel method to enhance local knowledge by using extensions of Local Context Network based on Proximity Values (LCPV) and Syntax-clusters Attention (SCA), named LCSA. LCPV first gets the induced trees from pre-trained models and generates the syntactic proximity values between context word and aspect to adaptively determine the extent of local context. Our improved SCA further extracts fine-grained knowledge, which not only focuses on the essential clusters for the target aspect term but also guides the model to learn essential words inside each cluster in DSPT. Extensive experimental results on multiple benchmark datasets demonstrate that LCSA is highly robust and achieves state-of-the-art performance for ALSA.

方面情感分析(ALSA)旨在提取句子中不同方面术语的极性。以前的研究利用传统的依赖语法分析树(DSPT)来编码上下文句法信息,取得了先进的成果。但是,这些研究可能无法有效地学习细粒度的句法知识,因此难以利用局部语境的优势。此外,这些研究也未能充分利用 DSPT 的依赖关系。为了解决这些问题,我们提出了一种新方法来增强本地知识,即使用基于邻近值的本地上下文网络(LCPV)和语法聚类注意力(SCA)的扩展,并将其命名为 LCSA。LCPV 首先从预先训练的模型中获取诱导树,然后生成上下文单词和方面之间的句法邻近值,从而自适应地确定本地上下文的范围。我们改进的 SCA 进一步提取了细粒度知识,这些知识不仅关注目标方面词的重要聚类,而且还引导模型学习 DSPT 中每个聚类内的重要词。在多个基准数据集上的广泛实验结果表明,LCSA 具有很强的鲁棒性,在 ALSA 方面达到了最先进的性能。
{"title":"Enhanced local knowledge with proximity values and syntax-clusters for aspect-level sentiment analysis","authors":"Pengfei Chen ,&nbsp;Biqing Zeng ,&nbsp;Yuwu Lu ,&nbsp;Yun Xue ,&nbsp;Fei Fan ,&nbsp;Mayi Xu ,&nbsp;Lingcong Feng","doi":"10.1016/j.csl.2023.101616","DOIUrl":"10.1016/j.csl.2023.101616","url":null,"abstract":"<div><p>Aspect-level sentiment analysis (ALSA) aims to extract the polarity of different aspect terms in a sentence. Previous works leveraging traditional dependency syntax parsing<span> trees (DSPT) to encode contextual syntactic<span> information had obtained state-of-the-art results. However, these works may not be able to learn fine-grained syntactic knowledge efficiently, which makes them difficult to take advantage of local context. Furthermore, these works failed to exploit the dependency relation from DSPT sufficiently. To solve these problems, we propose a novel method to enhance local knowledge by using extensions of Local Context Network based on Proximity Values (LCPV) and Syntax-clusters Attention (SCA), named LCSA. LCPV first gets the induced trees from pre-trained models and generates the syntactic proximity values between context word and aspect to adaptively determine the extent of local context. Our improved SCA further extracts fine-grained knowledge, which not only focuses on the essential clusters for the target aspect term but also guides the model to learn essential words inside each cluster in DSPT. Extensive experimental results on multiple benchmark datasets demonstrate that LCSA is highly robust and achieves state-of-the-art performance for ALSA.</span></span></p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2023-12-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139071559","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing accuracy and privacy in speech-based depression detection through speaker disentanglement 通过分离说话者提高基于语音的抑郁检测的准确性和私密性
IF 4.3 3区 计算机科学 Q1 Mathematics Pub Date : 2023-12-26 DOI: 10.1016/j.csl.2023.101605
Vijay Ravi , Jinhan Wang , Jonathan Flint , Abeer Alwan

Speech signals are valuable biomarkers for assessing an individual’s mental health, including identifying Major Depressive Disorder (MDD) automatically. A frequently used approach in this regard is to employ features related to speaker identity, such as speaker-embeddings. However, over-reliance on speaker identity features in mental health screening systems can compromise patient privacy. Moreover, some aspects of speaker identity may not be relevant for depression detection and could serve as a bias factor that hampers system performance. To overcome these limitations, we propose disentangling speaker-identity information from depression-related information. Specifically, we present four distinct disentanglement methods to achieve this — adversarial speaker identification (SID)-loss maximization (ADV), SID-loss equalization with variance (LEV), SID-loss equalization using Cross-Entropy (LECE) and SID-loss equalization using KL divergence (LEKLD). Our experiments, which incorporated diverse input features and model architectures, have yielded improved F1 scores for MDD detection and voice-privacy attributes, as quantified by Gain in Voice Distinctiveness (GVD) and De-Identification Scores (DeID). On the DAIC-WOZ dataset (English), LECE using ComparE16 features results in the best F1-Scores of 80% which represents the audio-only SOTA depression detection F1-Score along with a GVD of −1.1 dB and a DeID of 85%. On the EATD dataset (Mandarin), ADV using raw-audio signal achieves an F1-Score of 72.38% surpassing multi-modal SOTA along with a GVD of −0.89 dB dB and a DeID of 51.21%. By reducing the dependence on speaker-identity-related features, our method offers a promising direction for speech-based depression detection that preserves patient privacy.

语音信号是评估个人心理健康的重要生物标记,包括自动识别重度抑郁症(MDD)。这方面常用的一种方法是采用与说话者身份相关的特征,如说话者嵌入。然而,在心理健康筛查系统中过度依赖说话者身份特征可能会损害病人的隐私。此外,说话者身份的某些方面可能与抑郁检测无关,可能成为影响系统性能的偏差因素。为了克服这些局限性,我们建议将说话者身份信息与抑郁相关信息分离开来。具体来说,我们提出了四种不同的解缠方法来实现这一目标--对抗性说话人识别(SID)-损失最大化(ADV)、带方差的 SID 损失均衡(LEV)、使用交叉熵的 SID 损失均衡(LECE)和使用 KL 分歧的 SID 损失均衡(LEKLD)。我们的实验采用了不同的输入特征和模型架构,提高了 MDD 检测和语音隐私属性的 F1 分数,并通过语音独特性增益(GVD)和去识别分数(DeID)进行量化。在 DAIC-WOZ 数据集(英语)上,使用 ComparE16 特征的 LECE 得到了 80% 的最佳 F1 分数,代表了纯音频 SOTA 抑郁症检测 F1 分数,同时 GVD 为 -1.1 dB,DeID 为 85%。在 EATD 数据集(普通话)上,使用原始音频信号的 ADV 的 F1 分数为 72.38%,超过了多模式 SOTA,GVD 为 -0.89 dB dB,DeID 为 51.21%。通过减少对说话者身份相关特征的依赖,我们的方法为基于语音的抑郁检测提供了一个保护患者隐私的前景广阔的方向。
{"title":"Enhancing accuracy and privacy in speech-based depression detection through speaker disentanglement","authors":"Vijay Ravi ,&nbsp;Jinhan Wang ,&nbsp;Jonathan Flint ,&nbsp;Abeer Alwan","doi":"10.1016/j.csl.2023.101605","DOIUrl":"10.1016/j.csl.2023.101605","url":null,"abstract":"<div><p>Speech signals are valuable biomarkers for assessing an individual’s mental health, including identifying Major Depressive Disorder (MDD) automatically. A frequently used approach in this regard is to employ features related to speaker identity, such as speaker-embeddings. However, over-reliance on speaker identity features in mental health screening systems can compromise patient privacy. Moreover, some aspects of speaker identity may not be relevant for depression detection and could serve as a bias factor that hampers system performance. To overcome these limitations, we propose disentangling speaker-identity information from depression-related information. Specifically, we present four distinct disentanglement methods to achieve this — adversarial speaker identification (SID)-loss maximization (ADV), SID-loss equalization with variance (LEV), SID-loss equalization using Cross-Entropy (LECE) and SID-loss equalization using KL divergence (LEKLD). Our experiments, which incorporated diverse input features and model architectures, have yielded improved F1 scores for MDD detection and voice-privacy attributes, as quantified by Gain in Voice Distinctiveness (<span><math><msub><mrow><mi>G</mi></mrow><mrow><mi>V</mi><mi>D</mi></mrow></msub></math></span>) and De-Identification Scores (DeID). On the DAIC-WOZ dataset (English), LECE using ComparE16 features results in the best F1-Scores of 80% which represents the audio-only SOTA depression detection F1-Score along with a <span><math><msub><mrow><mi>G</mi></mrow><mrow><mi>V</mi><mi>D</mi></mrow></msub></math></span> of −1.1 dB and a DeID of 85%. On the EATD dataset (Mandarin), ADV using raw-audio signal achieves an F1-Score of 72.38% surpassing multi-modal SOTA along with a <span><math><msub><mrow><mi>G</mi></mrow><mrow><mi>V</mi><mi>D</mi></mrow></msub></math></span> of −0.89 dB dB and a DeID of 51.21%. By reducing the dependence on speaker-identity-related features, our method offers a promising direction for speech-based depression detection that preserves patient privacy.</p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2023-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0885230823001249/pdfft?md5=7acff7dbe3c70a9a6ae6cde978bd02e2&pid=1-s2.0-S0885230823001249-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139052205","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SecNLP: An NLP classification model watermarking framework based on multi-task learning SecNLP:基于多任务学习的 NLP 分类模型水印框架
IF 4.3 3区 计算机科学 Q1 Mathematics Pub Date : 2023-12-23 DOI: 10.1016/j.csl.2023.101606
Long Dai, Jiarong Mao, Liaoran Xu, Xuefeng Fan, Xiaoyi Zhou

The popularity of ChatGPT demonstrates the immense commercial value of natural language processing (NLP) technology. However, NLP models like ChatGPT are vulnerable to piracy and redistribution, which can harm the economic interests of model owners. Existing NLP model watermarking schemes struggle to balance robustness and covertness. Typically, robust watermarks require embedding more information, which compromises their covertness; conversely, covert watermarks are challenging to embed more information, which affects their robustness. This paper is proposed to use multi-task learning (MTL) to address the conflict between robustness and covertness. Specifically, a covert trigger set is established to implement remote verification of the watermark model, and a covert auxiliary network is designed to enhance the watermark model’s robustness. The proposed watermarking framework is evaluated on two benchmark datasets and three mainstream NLP models. Compared with existing schemes, the framework not only has excellent covertness and robustness but also has a lower false positive rate and can effectively resist fraudulent ownership claims by adversaries.

ChatGPT 的流行证明了自然语言处理(NLP)技术的巨大商业价值。然而,像 ChatGPT 这样的 NLP 模型很容易受到盗版和再传播的影响,从而损害模型所有者的经济利益。现有的 NLP 模型水印方案很难在稳健性和隐蔽性之间取得平衡。通常情况下,稳健型水印需要嵌入更多信息,这就会影响其隐蔽性;反之,隐蔽型水印则需要嵌入更多信息,这就会影响其稳健性。本文提出利用多任务学习(MTL)来解决稳健性和隐蔽性之间的矛盾。具体来说,建立隐蔽触发集来实现水印模型的远程验证,设计隐蔽辅助网络来增强水印模型的鲁棒性。我们在两个基准数据集和三个主流 NLP 模型上对所提出的水印框架进行了评估。与现有方案相比,该框架不仅具有出色的隐蔽性和鲁棒性,而且误报率较低,能有效抵御对手的所有权欺诈要求。
{"title":"SecNLP: An NLP classification model watermarking framework based on multi-task learning","authors":"Long Dai,&nbsp;Jiarong Mao,&nbsp;Liaoran Xu,&nbsp;Xuefeng Fan,&nbsp;Xiaoyi Zhou","doi":"10.1016/j.csl.2023.101606","DOIUrl":"10.1016/j.csl.2023.101606","url":null,"abstract":"<div><p><span>The popularity of ChatGPT demonstrates the immense commercial value of natural language processing (NLP) technology. However, NLP models like ChatGPT are vulnerable to piracy and redistribution, which can harm the economic interests of model owners. Existing NLP model </span>watermarking schemes<span> struggle to balance robustness and covertness. Typically, robust watermarks require embedding more information, which compromises their covertness; conversely, covert watermarks are challenging to embed more information, which affects their robustness. This paper is proposed to use multi-task learning (MTL) to address the conflict between robustness and covertness. Specifically, a covert trigger set is established to implement remote verification of the watermark model, and a covert auxiliary network is designed to enhance the watermark model’s robustness. The proposed watermarking framework is evaluated on two benchmark datasets and three mainstream NLP models. Compared with existing schemes, the framework not only has excellent covertness and robustness but also has a lower false positive rate and can effectively resist fraudulent ownership claims by adversaries.</span></p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2023-12-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139030523","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Contextual emotion detection using ensemble deep learning 利用集合深度学习进行情境情感检测
IF 4.3 3区 计算机科学 Q1 Mathematics Pub Date : 2023-12-20 DOI: 10.1016/j.csl.2023.101604
Asalah Thiab , Luay Alawneh , Mohammad AL-Smadi

Emotion detection from online textual information is gaining more attention due to its usefulness in understanding users’ behaviors and their desires. This is driven by the large amounts of texts from different sources such as social media and shopping websites. Recent studies investigated the benefits of deep learning in the detection of emotions from textual conversations. In this paper, we study the performance of several deep learning and transformer-based models in the classification of emotions in English conversations. Further, we apply ensemble learning using a majority voting technique to improve the overall classification performance. We evaluated our proposed models on the SemEval 2019 Task 3 public dataset that categorizes emotions as Happy, Angry, Sad, and Others. The results show that our models can successfully distinguish the three main classes of emotions and separate them from Others in a highly imbalanced dataset. The transformer-based models achieved a micro-averaged F1-score of up to 75.55%, whereas the RNN-based models only reached 67.03%. Further, we show that the ensemble model significantly improves the overall performance and achieves a micro-averaged F1-score of 77.07%.

从在线文本信息中进行情感检测有助于了解用户的行为和愿望,因此越来越受到人们的关注。这主要得益于来自社交媒体和购物网站等不同来源的大量文本。最近的研究调查了深度学习在从文本对话中检测情感方面的优势。在本文中,我们研究了几种基于深度学习和转换器的模型在英语会话情感分类中的表现。此外,我们还利用多数投票技术进行了集合学习,以提高整体分类性能。我们在 SemEval 2019 Task 3 公共数据集上评估了我们提出的模型,该数据集将情绪分类为快乐、愤怒、悲伤和其他。结果表明,我们的模型可以成功区分三大类情绪,并在高度不平衡的数据集中将它们与 "其他 "区分开来。基于变换器的模型的微平均 F1 分数高达 75.55%,而基于 RNN 的模型仅为 67.03%。此外,我们还发现,集合模型显著提高了整体性能,微平均 F1 分数达到了 77.07%。
{"title":"Contextual emotion detection using ensemble deep learning","authors":"Asalah Thiab ,&nbsp;Luay Alawneh ,&nbsp;Mohammad AL-Smadi","doi":"10.1016/j.csl.2023.101604","DOIUrl":"10.1016/j.csl.2023.101604","url":null,"abstract":"<div><p><span><span>Emotion detection from online textual information is gaining more attention due to its usefulness in understanding users’ behaviors and their desires. This is driven by the large amounts of texts from different sources such as social media and shopping websites. Recent studies investigated the benefits of deep learning in the detection of emotions from textual conversations. In this paper, we study the performance of several deep learning and transformer-based models in the classification of emotions in English conversations. Further, we apply </span>ensemble learning using a majority voting technique to improve the overall classification performance. We evaluated our proposed models on the SemEval 2019 Task 3 public dataset that categorizes emotions as </span><em>Happy</em>, <em>Angry</em>, <em>Sad</em>, and <em>Others</em>. The results show that our models can successfully distinguish the three main classes of emotions and separate them from <em>Others</em> in a highly imbalanced dataset. The transformer-based models achieved a micro-averaged F1-score of up to 75.55%, whereas the RNN-based models only reached 67.03%. Further, we show that the ensemble model significantly improves the overall performance and achieves a micro-averaged F1-score of 77.07%.</p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2023-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139030524","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving self-supervised learning model for audio spoofing detection with layer-conditioned embedding fusion 利用层条件嵌入融合改进音频欺骗检测的自监督学习模型
IF 4.3 3区 计算机科学 Q1 Mathematics Pub Date : 2023-12-18 DOI: 10.1016/j.csl.2023.101599
Souvik Sinha, Spandan Dey, Goutam Saha

The application of voice recognition systems has increased by a great deal with technology. This has allowed adversaries to falsely claim access to these systems by spoofing the identity of a target speaker. The existing supervised learning (SL)-based countermeasures are yet to provide a complete solution against the newly evolving spoofing attacks. To tackle this problem, we explore self-supervised learning (SSL)-based frameworks. At first, we implement widely used SSL frameworks, where our target is identifying spoofed speech. We report a considerable performance improvement over the SL state-of-the-art baseline as a whole. Then, we perform an attack-wise comparative analysis between SL and SSL frameworks. While the SSL performs better in most cases, there are certain attacks where the SL outperforms it. Hence, we hypothesize that there is scope to jointly utilize information effectively included by both these models for better performance. To do that, we first perform conventional weighted score fusion between the SL and best-performing SSL models, which reduces the EER, outperforming both the state-of-the-art SL and best-performing SSL framework. Then, we propose an embedding fusion scheme that minimizes the embedding distribution between the selected SL and SSL representations. To select the appropriate layers, we perform a comprehensive statistical analysis. The proposed fusion scheme outperforms the score fusion method and shows that the SSL performance can be improved by effectively including learned knowledge from the SL framework. The final EER achieved on the ASVspoof 2019 logical access (LA) database is 0.177%, a significant improvement over our baseline. Using the ASVspoof 2021 LA as a blind evaluation dataset, our proposed embedding fusion scheme reduces the EER to 2.666%.

随着技术的发展,语音识别系统的应用大大增加。这就使得对手可以通过欺骗目标说话者的身份,谎称可以访问这些系统。现有的基于监督学习(SL)的反制措施还不能提供完整的解决方案来应对新发展的欺骗攻击。为了解决这个问题,我们探索了基于自监督学习(SSL)的框架。首先,我们实施了广泛使用的 SSL 框架,目标是识别欺骗性语音。我们的报告显示,与基于自监督学习的最先进基准相比,我们的整体性能有了相当大的提高。然后,我们对 SL 和 SSL 框架进行了攻击方面的比较分析。虽然 SSL 在大多数情况下表现更好,但在某些攻击中,SL 的表现要优于 SSL。因此,我们假设可以联合利用这两种模型所包含的有效信息,以获得更好的性能。为此,我们首先在 SL 模型和表现最佳的 SSL 模型之间进行了传统的加权分数融合,从而降低了 EER,表现优于最先进的 SL 框架和表现最佳的 SSL 框架。然后,我们提出了一种嵌入式融合方案,它能使所选 SL 和 SSL 表示之间的嵌入分布最小化。为了选择合适的层,我们进行了全面的统计分析。所提出的融合方案优于分数融合方法,并表明通过有效纳入从 SL 框架中学到的知识,可以提高 SSL 性能。在 ASVspoof 2019 逻辑访问(LA)数据库上实现的最终 EER 为 0.177%,比我们的基线有显著提高。使用 ASVspoof 2021 LA 作为盲评估数据集,我们提出的嵌入融合方案将 EER 降低到 2.666%。
{"title":"Improving self-supervised learning model for audio spoofing detection with layer-conditioned embedding fusion","authors":"Souvik Sinha,&nbsp;Spandan Dey,&nbsp;Goutam Saha","doi":"10.1016/j.csl.2023.101599","DOIUrl":"10.1016/j.csl.2023.101599","url":null,"abstract":"<div><p>The application of voice recognition<span><span> systems has increased by a great deal with technology. This has allowed adversaries to falsely claim access to these systems by spoofing the identity of a target speaker. The existing supervised learning (SL)-based countermeasures<span> are yet to provide a complete solution against the newly evolving spoofing attacks. To tackle this problem, we explore self-supervised learning (SSL)-based frameworks. At first, we implement widely used SSL frameworks, where our target is identifying spoofed speech. We report a considerable performance improvement over the SL state-of-the-art baseline as a whole. Then, we perform an attack-wise comparative analysis between SL and SSL frameworks. While the SSL performs better in most cases, there are certain attacks where the SL outperforms it. Hence, we hypothesize that there is scope to jointly utilize information effectively included by both these models for better performance. To do that, we first perform conventional weighted score fusion between the SL and best-performing SSL models, which reduces the </span></span>EER, outperforming both the state-of-the-art SL and best-performing SSL framework. Then, we propose an embedding fusion scheme that minimizes the embedding distribution between the selected SL and SSL representations. To select the appropriate layers, we perform a comprehensive statistical analysis. The proposed fusion scheme outperforms the score fusion method and shows that the SSL performance can be improved by effectively including learned knowledge from the SL framework. The final EER achieved on the ASVspoof 2019 logical access (LA) database is 0.177%, a significant improvement over our baseline. Using the ASVspoof 2021 LA as a blind evaluation dataset, our proposed embedding fusion scheme reduces the EER to 2.666%.</span></p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2023-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138746029","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A novel channel estimate for noise robust speech recognition 用于鲁棒噪声语音识别的新型信道估计
IF 4.3 3区 计算机科学 Q1 Mathematics Pub Date : 2023-12-16 DOI: 10.1016/j.csl.2023.101598
Geoffroy Vanderreydt, Kris Demuynck

We propose a novel technique to estimate the channel characteristics for robust speech recognition. The method focuses on reliable time–frequency speech patches which are highly independent of the noise condition. Combined with a root-based approximation of the logarithm in the MFCC computation, this reduces the variance caused by the noise on the spectral features, and therefore also the constrain on the acoustic model in a multi-style training setup. We show that compared to the standard mean normalization, the proposed method estimates the channel equally well under clean conditions and better under noisy conditions. When integrated in the feature extraction pipeline, we show improvements in speech recognition accuracy on noisy speech and a status quo on clean speech. Our experiments reveal that this method helps the most for generative models that need to model the complex noise variability, and less so for discriminative models, which can learn to ignore noise instead of accurately modeling it. Our approach outperforms the state of the art on the noisy Aurora4 task.

我们提出了一种估算信道特征的新技术,以实现鲁棒语音识别。该方法侧重于与噪声条件高度无关的可靠时频语音片段。结合 MFCC 计算中基于根的对数近似,这就减少了噪声对频谱特征造成的方差,从而也减少了多风格训练设置中对声学模型的限制。我们的研究表明,与标准平均值归一化方法相比,所提出的方法在干净条件下对信道的估计效果相当好,而在噪声条件下则更好。当集成到特征提取管道中时,我们发现噪声语音的语音识别准确率有所提高,而干净语音的识别准确率则维持现状。我们的实验表明,这种方法对生成模型的帮助最大,因为生成模型需要对复杂的噪声变化进行建模,而对判别模型的帮助较小,因为判别模型可以学习忽略噪声,而不是对其进行精确建模。在噪声 Aurora4 任务中,我们的方法优于现有技术。
{"title":"A novel channel estimate for noise robust speech recognition","authors":"Geoffroy Vanderreydt,&nbsp;Kris Demuynck","doi":"10.1016/j.csl.2023.101598","DOIUrl":"10.1016/j.csl.2023.101598","url":null,"abstract":"<div><p>We propose a novel technique to estimate the channel characteristics for robust speech recognition<span>. The method focuses on reliable time–frequency speech patches which are highly independent of the noise condition. Combined with a root-based approximation<span> of the logarithm in the MFCC computation, this reduces the variance caused by the noise on the spectral features<span>, and therefore also the constrain on the acoustic model in a multi-style training setup. We show that compared to the standard mean normalization, the proposed method estimates the channel equally well under clean conditions and better under noisy conditions. When integrated in the feature extraction pipeline, we show improvements in speech recognition accuracy on noisy speech and a status quo on clean speech. Our experiments reveal that this method helps the most for generative models that need to model the complex noise variability, and less so for discriminative models, which can learn to ignore noise instead of accurately modeling it. Our approach outperforms the state of the art on the noisy Aurora4 task.</span></span></span></p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2023-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138745942","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computer Speech and Language
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1