首页 > 最新文献

IEEE Journal of Selected Topics in Signal Processing最新文献

英文 中文
Spatial-Spectral Mixing Transformer With Hybrid Image Prior for Multispectral Image Demosaicing 基于混合图像先验的空间光谱混合变压器多光谱图像去马赛克
IF 8.7 1区 工程技术 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-12 DOI: 10.1109/JSTSP.2024.3516374
Le Dong;Mengzu Liu;Tengteng Tang;Tao Huang;Jie Lin;Weisheng Dong;Guangming Shi
The snapshot multispectral imaging system using a multispectral filter array (MSFA) efficiently captures sample the multispectral image (MSI) information of scenes and obtain spectral mosaic images. To obtain the complete MSI information from these spectral mosaics, effective demosaicing methods are essential. Traditional MSI demosaicing techniques depend on pixel correlation and various hand-crafted priors, while deep learning-based approaches learn the mapping between spectral mosaic images and MSI directly. However, current methods often fall short in recovery performance, leaving significant room for improvement in the MSI demosaicing field. In this paper, we propose a novel MSI demosaicing method based on the spatial-spectral mixing transformer with hybrid image prior, named SSMT-HIP, to enhance image reconstruction and detail recovery. Our framework, the spatial-spectral mixing transformer (SSMT), is designed to comprehensively learn the spatial-spectral correlations of the data, addressing the limitations of current CNN-based methods in capturing both spatial and spectral characteristics of MSI. Furthermore, we introduce the deep hybrid image prior (HIP), which combines the deep Gaussian scale mixture (GSM) prior and the deep nonlocal auto-regressive (NAR) prior. This hybrid prior is learned in an end-to-end manner through the deep unfolding network. The GSM prior excels at recovering image textures and details, while the NAR prior enhances long-range dependencies in MSI. Extensive experiments on both synthetic and real-world data demonstrate that our proposed method outperforms existing state-of-the-art MSI demosaicing methods.
采用多光谱滤波阵列(MSFA)的快照多光谱成像系统可以有效地采集场景的多光谱图像信息,获得光谱拼接图像。为了从这些光谱马赛克中获得完整的MSI信息,有效的反马赛克方法是必不可少的。传统的MSI去马赛克技术依赖于像素相关和各种手工制作的先验,而基于深度学习的方法直接学习光谱马赛克图像与MSI之间的映射。然而,目前的方法在恢复性能上往往存在不足,因此在MSI反马赛克领域还有很大的改进空间。本文提出了一种基于混合图像先验的空间-光谱混合变换的MSI去马赛克方法,命名为SSMT-HIP,以增强图像重建和细节恢复。我们的框架,即空间-光谱混合变压器(SSMT),旨在全面学习数据的空间-光谱相关性,解决当前基于cnn的方法在捕获MSI的空间和光谱特征方面的局限性。此外,我们还引入了深度混合图像先验(HIP),它结合了深度高斯尺度混合先验(GSM)和深度非局部自回归先验(NAR)。这种混合先验通过深度展开网络以端到端方式学习。GSM先验在恢复图像纹理和细节方面表现出色,而NAR先验在MSI中增强了远程依赖性。在合成数据和实际数据上进行的大量实验表明,我们提出的方法优于现有的最先进的MSI反马赛克方法。
{"title":"Spatial-Spectral Mixing Transformer With Hybrid Image Prior for Multispectral Image Demosaicing","authors":"Le Dong;Mengzu Liu;Tengteng Tang;Tao Huang;Jie Lin;Weisheng Dong;Guangming Shi","doi":"10.1109/JSTSP.2024.3516374","DOIUrl":"https://doi.org/10.1109/JSTSP.2024.3516374","url":null,"abstract":"The snapshot multispectral imaging system using a multispectral filter array (MSFA) efficiently captures sample the multispectral image (MSI) information of scenes and obtain spectral mosaic images. To obtain the complete MSI information from these spectral mosaics, effective demosaicing methods are essential. Traditional MSI demosaicing techniques depend on pixel correlation and various hand-crafted priors, while deep learning-based approaches learn the mapping between spectral mosaic images and MSI directly. However, current methods often fall short in recovery performance, leaving significant room for improvement in the MSI demosaicing field. In this paper, we propose a novel MSI demosaicing method based on the spatial-spectral mixing transformer with hybrid image prior, named SSMT-HIP, to enhance image reconstruction and detail recovery. Our framework, the spatial-spectral mixing transformer (SSMT), is designed to comprehensively learn the spatial-spectral correlations of the data, addressing the limitations of current CNN-based methods in capturing both spatial and spectral characteristics of MSI. Furthermore, we introduce the deep hybrid image prior (HIP), which combines the deep Gaussian scale mixture (GSM) prior and the deep nonlocal auto-regressive (NAR) prior. This hybrid prior is learned in an end-to-end manner through the deep unfolding network. The GSM prior excels at recovering image textures and details, while the NAR prior enhances long-range dependencies in MSI. Extensive experiments on both synthetic and real-world data demonstrate that our proposed method outperforms existing state-of-the-art MSI demosaicing methods.","PeriodicalId":13038,"journal":{"name":"IEEE Journal of Selected Topics in Signal Processing","volume":"19 1","pages":"221-233"},"PeriodicalIF":8.7,"publicationDate":"2024-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143512870","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Vision Language Modeling of Content, Distortion and Appearance for Image Quality Assessment 用于图像质量评估的内容、失真和外观视觉语言建模
IF 8.7 1区 工程技术 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-12 DOI: 10.1109/JSTSP.2024.3516392
Fei Zhou;Tianhao Gu;Zhicong Huang;Guoping Qiu
The visual quality of an image is confounded by a number of intertwined factors including its semantic content, distortion characteristics and appearance properties such as brightness, contrast, sharpness, and colourfulness. Distilling high level knowledge about all these quality bearing attributes is crucial for developing objective Image Quality Assessment (IQA). While existing solutions have modeled some of these aspects, a comprehensive solution that involves all these important quality related attributes has not yet been developed. In this paper, we present a new blind IQA (BIQA) model termed Self-supervision and Vision-Language supervision Image QUality Evaluator (SLIQUE) that features a joint vision-language and visual contrastive representation learning framework for acquiring high level knowledge about the images semantic contents, distortion characteristics and appearance properties for IQA. For training SLIQUE, we have developed a systematic approach to constructing a first of its kind large image database annotated with all three categories of quality relevant texts. The Text Annotated Distortion, Appearance and Content (TADAC1) database has over 1.6 million images annotated with textual descriptions of their semantic contents, distortion characteristics and appearance properties. The method for constructing TADAC and the database itself will be particularly useful for exploiting vision-language modeling for advanced IQA applications. Extensive experimental results show that SLIQUE has superior performances over state of the art, demonstrating the soundness of its design principle and the effectiveness of its implementation.
图像的视觉质量受到许多相互交织的因素的影响,包括其语义内容、失真特征和外观属性,如亮度、对比度、清晰度和色彩。提取关于所有这些质量承载属性的高层次知识对于开发客观的图像质量评估(IQA)至关重要。虽然现有的解决方案已经对其中的一些方面进行了建模,但是还没有开发出包含所有这些重要的质量相关属性的综合解决方案。在本文中,我们提出了一个新的盲IQA (BIQA)模型,称为自监督和视觉语言监督图像质量评估器(SLIQUE),该模型具有视觉语言和视觉对比表征的联合学习框架,用于获取关于IQA图像语义内容、失真特征和外观属性的高级知识。为了训练SLIQUE,我们开发了一种系统的方法来构建第一个带有所有三类高质量相关文本注释的大型图像数据库。文本标注畸变、外观和内容(TADAC1)数据库拥有超过160万张图像,并对其语义内容、畸变特征和外观属性进行了文本描述。构建TADAC和数据库本身的方法对于开发高级IQA应用程序的视觉语言建模特别有用。大量的实验结果表明,SLIQUE具有优于现有技术的性能,证明了其设计原理的合理性和实施的有效性。
{"title":"Vision Language Modeling of Content, Distortion and Appearance for Image Quality Assessment","authors":"Fei Zhou;Tianhao Gu;Zhicong Huang;Guoping Qiu","doi":"10.1109/JSTSP.2024.3516392","DOIUrl":"https://doi.org/10.1109/JSTSP.2024.3516392","url":null,"abstract":"The visual quality of an image is confounded by a number of intertwined factors including its semantic content, distortion characteristics and appearance properties such as brightness, contrast, sharpness, and colourfulness. Distilling high level knowledge about all these quality bearing attributes is crucial for developing objective Image Quality Assessment (IQA). While existing solutions have modeled some of these aspects, a comprehensive solution that involves all these important quality related attributes has not yet been developed. In this paper, we present a new blind IQA (BIQA) model termed Self-supervision and Vision-Language supervision Image QUality Evaluator (SLIQUE) that features a joint vision-language and visual contrastive representation learning framework for acquiring high level knowledge about the images semantic contents, distortion characteristics and appearance properties for IQA. For training SLIQUE, we have developed a systematic approach to constructing a first of its kind large image database annotated with all three categories of quality relevant texts. The Text Annotated Distortion, Appearance and Content (TADAC<sup>1</sup>) database has over 1.6 million images annotated with textual descriptions of their semantic contents, distortion characteristics and appearance properties. The method for constructing TADAC and the database itself will be particularly useful for exploiting vision-language modeling for advanced IQA applications. Extensive experimental results show that SLIQUE has superior performances over state of the art, demonstrating the soundness of its design principle and the effectiveness of its implementation.","PeriodicalId":13038,"journal":{"name":"IEEE Journal of Selected Topics in Signal Processing","volume":"19 1","pages":"234-247"},"PeriodicalIF":8.7,"publicationDate":"2024-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143512839","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Signal Processing and Learning for Next Generation Multiple Access in 6G 下一代6G多址的信号处理与学习
IF 8.7 1区 工程技术 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-09 DOI: 10.1109/JSTSP.2024.3511403
Wei Chen;Yuanwei Liu;Hamid Jafarkhani;Yonina C. Eldar;Peiying Zhu;Khaled B. Letaief
Wireless communication systems to date primarily rely on the orthogonality of resources to facilitate the design and implementation, from user access to data transmission. Emerging applications and scenarios in the sixth generation (6G) wireless systems will require massive connectivity and transmission of a deluge of data, which calls for more flexibility in the design concept that goes beyond orthogonality. Furthermore, recent advances in signal processing and learning, e.g., deep learning, provide promising approaches to deal with complex and previously intractable problems. This article provides an overview of research efforts to date in the field of signal processing and learning for next-generation multiple access, with an emphasis on massive random access and non-orthogonal multiple access. The promising interplay with new technologies and the challenges in learning-based NGMA are discussed.
迄今为止,无线通信系统主要依靠资源的正交性来方便设计和实现,从用户访问到数据传输。第六代(6G)无线系统中的新兴应用和场景将需要大规模连接和传输海量数据,这就要求在设计概念上具有超越正交性的更大灵活性。此外,信号处理和学习的最新进展,例如深度学习,为处理复杂和以前棘手的问题提供了有前途的方法。本文概述了迄今为止在下一代多址信号处理和学习领域的研究成果,重点介绍了大规模随机接入和非正交多址接入。讨论了基于学习的NGMA与新技术的相互作用以及面临的挑战。
{"title":"Signal Processing and Learning for Next Generation Multiple Access in 6G","authors":"Wei Chen;Yuanwei Liu;Hamid Jafarkhani;Yonina C. Eldar;Peiying Zhu;Khaled B. Letaief","doi":"10.1109/JSTSP.2024.3511403","DOIUrl":"https://doi.org/10.1109/JSTSP.2024.3511403","url":null,"abstract":"Wireless communication systems to date primarily rely on the orthogonality of resources to facilitate the design and implementation, from user access to data transmission. Emerging applications and scenarios in the sixth generation (6G) wireless systems will require massive connectivity and transmission of a deluge of data, which calls for more flexibility in the design concept that goes beyond orthogonality. Furthermore, recent advances in signal processing and learning, e.g., deep learning, provide promising approaches to deal with complex and previously intractable problems. This article provides an overview of research efforts to date in the field of signal processing and learning for next-generation multiple access, with an emphasis on massive random access and non-orthogonal multiple access. The promising interplay with new technologies and the challenges in learning-based NGMA are discussed.","PeriodicalId":13038,"journal":{"name":"IEEE Journal of Selected Topics in Signal Processing","volume":"18 7","pages":"1146-1177"},"PeriodicalIF":8.7,"publicationDate":"2024-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142993322","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Gradient-Free Post-Hoc Explainability Using Distillation Aided Learnable Approach 使用蒸馏辅助可学习方法的无梯度事后解释性
IF 8.7 1区 工程技术 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-04 DOI: 10.1109/JSTSP.2024.3467914
Debarpan Bhattacharya;Amir H. Poorjam;Deepak Mittal;Sriram Ganapathy
The recent advancements in artificial intelligence (AI), with the release of several large models having only query access, make a strong case for explainability of deep models in a post-hoc gradient free manner. In this paper, we propose a framework, named distillation aided explainability (DAX), that attempts to generate a saliency-based explanation in a model agnostic gradient free application. The DAX approach poses the problem of explanation in a learnable setting with a mask generation network and a distillation network. The mask generation network learns to generate the multiplier mask that finds the salient regions of the input, while the student distillation network aims to approximate the local behavior of the black-box model. We propose a joint optimization of the two networks in the DAX framework using the locally perturbed input samples, with the targets derived from input-output access to the black-box model. We extensively evaluate DAX across different modalities (image and audio), in a classification setting, using a diverse set of evaluations (intersection over union with ground truth, deletion based and subjective human evaluation based measures) and benchmark it with respect to 9 different methods. In these evaluations, the DAX significantly outperforms the existing approaches on all modalities and evaluation metrics.
人工智能(AI)最近的进步,随着几个只有查询访问权限的大型模型的发布,以一种事后梯度自由的方式对深度模型的可解释性进行了强有力的论证。在本文中,我们提出了一个名为蒸馏辅助解释性(DAX)的框架,它试图在模型不可知的梯度自由应用程序中生成基于显著性的解释。DAX方法提出了在可学习设置下的解释问题,包括掩模生成网络和蒸馏网络。掩码生成网络学习生成找到输入显著区域的乘法器掩码,而学生蒸馏网络旨在近似黑箱模型的局部行为。我们提出了在DAX框架中使用局部摄动输入样本对两个网络进行联合优化,目标来源于对黑盒模型的输入输出访问。我们在不同的模式(图像和音频)中广泛评估DAX,在分类设置中,使用不同的评估集(与基础事实的交叉结合,基于删除和基于主观人类评估的措施),并根据9种不同的方法对其进行基准测试。在这些评估中,DAX在所有模式和评估指标上都明显优于现有方法。
{"title":"Gradient-Free Post-Hoc Explainability Using Distillation Aided Learnable Approach","authors":"Debarpan Bhattacharya;Amir H. Poorjam;Deepak Mittal;Sriram Ganapathy","doi":"10.1109/JSTSP.2024.3467914","DOIUrl":"https://doi.org/10.1109/JSTSP.2024.3467914","url":null,"abstract":"The recent advancements in artificial intelligence (AI), with the release of several large models having only query access, make a strong case for explainability of deep models in a post-hoc gradient free manner. In this paper, we propose a framework, named distillation aided explainability (DAX), that attempts to generate a saliency-based explanation in a model agnostic gradient free application. The DAX approach poses the problem of explanation in a learnable setting with a mask generation network and a distillation network. The mask generation network learns to generate the multiplier mask that finds the salient regions of the input, while the student distillation network aims to approximate the local behavior of the black-box model. We propose a joint optimization of the two networks in the DAX framework using the locally perturbed input samples, with the targets derived from input-output access to the black-box model. We extensively evaluate DAX across different modalities (image and audio), in a classification setting, using a diverse set of evaluations (intersection over union with ground truth, deletion based and subjective human evaluation based measures) and benchmark it with respect to 9 different methods. In these evaluations, the DAX significantly outperforms the existing approaches on all modalities and evaluation metrics.","PeriodicalId":13038,"journal":{"name":"IEEE Journal of Selected Topics in Signal Processing","volume":"19 1","pages":"169-180"},"PeriodicalIF":8.7,"publicationDate":"2024-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143512982","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Representation Transfer Learning via Multiple Pre-Trained Models for Linear Regression 基于多个预训练模型的线性回归表征迁移学习
IF 8.7 1区 工程技术 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-11-27 DOI: 10.1109/JSTSP.2024.3507371
Navjot Singh;Suhas Diggavi
In this paper, we consider the problem of learning a linear regression model on a data domain of interest (target) given few samples. To aid learning, we are provided with a set of pre-trained regression models that are trained on potentially different data domains (sources). Assuming a representation structure for the data generating linear models at the sources and the target domains, we propose a representation transfer based learning method for constructing the target model. The proposed scheme is comprised of two phases: (i) utilizing the different source representations to construct a representation that is adapted to the target data, and (ii) using the obtained model as an initialization to a fine-tuning procedure that re-trains the entire (over-parameterized) regression model on the target data. For each phase of the training method, we provide excess risk bounds for the learned model compared to the true data generating target model. The derived bounds show a gain in sample complexity for our proposed method compared to the baseline method of not leveraging source representations when achieving the same excess risk, therefore, theoretically demonstrating the effectiveness of transfer learning for linear regression.
在本文中,我们考虑在给定少量样本的感兴趣数据域(目标)上学习线性回归模型的问题。为了帮助学习,我们提供了一组预训练的回归模型,这些模型是在可能不同的数据域(来源)上训练的。假设在源域和目标域生成线性模型的数据具有一定的表示结构,提出了一种基于表示迁移的学习方法来构建目标模型。提出的方案由两个阶段组成:(i)利用不同的源表示来构建适应目标数据的表示,以及(ii)使用获得的模型作为微调过程的初始化,该过程重新训练目标数据上的整个(过度参数化)回归模型。对于训练方法的每个阶段,我们为学习模型提供了与真实数据生成目标模型相比的超额风险界限。与不利用源表示的基线方法相比,导出的边界显示了我们提出的方法在样本复杂性方面的增益,因此,从理论上证明了线性回归迁移学习的有效性。
{"title":"Representation Transfer Learning via Multiple Pre-Trained Models for Linear Regression","authors":"Navjot Singh;Suhas Diggavi","doi":"10.1109/JSTSP.2024.3507371","DOIUrl":"https://doi.org/10.1109/JSTSP.2024.3507371","url":null,"abstract":"In this paper, we consider the problem of learning a linear regression model on a data domain of interest (target) given few samples. To aid learning, we are provided with a set of pre-trained regression models that are trained on potentially different data domains (sources). Assuming a representation structure for the data generating linear models at the sources and the target domains, we propose a representation transfer based learning method for constructing the target model. The proposed scheme is comprised of two phases: (i) utilizing the different source representations to construct a representation that is adapted to the target data, and (ii) using the obtained model as an initialization to a fine-tuning procedure that re-trains the entire (over-parameterized) regression model on the target data. For each phase of the training method, we provide excess risk bounds for the learned model compared to the true data generating target model. The derived bounds show a gain in sample complexity for our proposed method compared to the baseline method of not leveraging source representations when achieving the same excess risk, therefore, theoretically demonstrating the effectiveness of transfer learning for linear regression.","PeriodicalId":13038,"journal":{"name":"IEEE Journal of Selected Topics in Signal Processing","volume":"19 1","pages":"208-220"},"PeriodicalIF":8.7,"publicationDate":"2024-11-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143512857","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SemantiCodec: An Ultra Low Bitrate Semantic Audio Codec for General Sound SemantiCodec:一个超低比特率的通用声音语义音频编解码器
IF 8.7 1区 工程技术 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-11-26 DOI: 10.1109/JSTSP.2024.3506286
Haohe Liu;Xuenan Xu;Yi Yuan;Mengyue Wu;Wenwu Wang;Mark D. Plumbley
Large languagemodels (LLMs) have significantly advanced audio processing through audio codecs that convert audio into discrete tokens, enabling the application of language modelling techniques to audio data. However, traditional codecs often operate at high bitrates or within narrow domains such as speech and lack the semantic clues required for efficient language modelling. Addressing these challenges, we introduce SemantiCodec, a novel codec designed to compress audio into fewer than a hundred tokens per second across diverse audio types, including speech, general sound, and music, without compromising quality. SemantiCodec features a dual-encoder architecture: a semantic encoder using a self-supervised pre-trained Audio Masked Autoencoder (AudioMAE), discretized using k-means clustering on extensive audio data, and an acoustic encoder to capture the remaining details. The semantic and acoustic encoder outputs are used to reconstruct audio via a diffusion-model-based decoder. SemantiCodec is presented in three variants with token rates of 25, 50, and 100 per second, supporting a range of ultra-low bit rates between 0.31 kbps and 1.40 kbps. Experimental results demonstrate that SemantiCodec significantly outperforms the state-of-the-art Descript codec on reconstruction quality. Our results also suggest that SemantiCodec contains significantly richer semantic information than all evaluated state-of-the-art audio codecs, even at significantly lower bitrates.
大型语言模型(LLM)通过音频编解码器将音频转换为离散的词块,从而使语言建模技术能够应用于音频数据,大大推进了音频处理技术的发展。然而,传统编解码器通常以高比特率或在语音等狭窄领域内运行,缺乏高效语言建模所需的语义线索。为了应对这些挑战,我们推出了 SemantiCodec,这是一种新颖的编解码器,旨在将各种音频类型(包括语音、一般声音和音乐)的音频压缩到每秒不到一百个词组,同时不影响质量。SemantiCodec 采用双编码器架构:语义编码器使用自监督预训练的音频屏蔽自动编码器(AudioMAE),通过对大量音频数据进行 k-means 聚类来离散化;声学编码器捕捉其余细节。语义编码器和声学编码器的输出用于通过基于扩散模型的解码器重建音频。SemantiCodec 有三个变体,标记率分别为每秒 25、50 和 100,支持 0.31 kbps 至 1.40 kbps 的超低比特率范围。实验结果表明,SemantiCodec 在重构质量上明显优于最先进的 Descript 编解码器。我们的结果还表明,SemantiCodec 包含的语义信息比所有经过评估的最先进音频编解码器要丰富得多,即使在比特率明显较低的情况下也是如此。
{"title":"SemantiCodec: An Ultra Low Bitrate Semantic Audio Codec for General Sound","authors":"Haohe Liu;Xuenan Xu;Yi Yuan;Mengyue Wu;Wenwu Wang;Mark D. Plumbley","doi":"10.1109/JSTSP.2024.3506286","DOIUrl":"https://doi.org/10.1109/JSTSP.2024.3506286","url":null,"abstract":"Large languagemodels (LLMs) have significantly advanced audio processing through audio codecs that convert audio into discrete tokens, enabling the application of language modelling techniques to audio data. However, traditional codecs often operate at high bitrates or within narrow domains such as speech and lack the semantic clues required for efficient language modelling. Addressing these challenges, we introduce SemantiCodec, a novel codec designed to compress audio into fewer than a hundred tokens per second across diverse audio types, including speech, general sound, and music, without compromising quality. SemantiCodec features a dual-encoder architecture: a semantic encoder using a self-supervised pre-trained Audio Masked Autoencoder (AudioMAE), discretized using k-means clustering on extensive audio data, and an acoustic encoder to capture the remaining details. The semantic and acoustic encoder outputs are used to reconstruct audio via a diffusion-model-based decoder. SemantiCodec is presented in three variants with token rates of 25, 50, and 100 per second, supporting a range of ultra-low bit rates between 0.31 kbps and 1.40 kbps. Experimental results demonstrate that SemantiCodec significantly outperforms the state-of-the-art Descript codec on reconstruction quality. Our results also suggest that SemantiCodec contains significantly richer semantic information than all evaluated state-of-the-art audio codecs, even at significantly lower bitrates.","PeriodicalId":13038,"journal":{"name":"IEEE Journal of Selected Topics in Signal Processing","volume":"18 8","pages":"1448-1461"},"PeriodicalIF":8.7,"publicationDate":"2024-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143184457","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Coding Speech Through Vocal Tract Kinematics 基于声道运动学的语音编码
IF 8.7 1区 工程技术 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-11-20 DOI: 10.1109/JSTSP.2024.3497655
Cheol Jun Cho;Peter Wu;Tejas S. Prabhune;Dhruv Agarwal;Gopala K. Anumanchipalli
Vocal tract articulation is a natural, grounded control space of speech production. The spatiotemporal coordination of articulators combined with the vocal source shapes intelligible speech sounds to enable effective spoken communication. Based on this physiological grounding of speech, we propose a new framework of neural encoding-decoding of speech – Speech Articulatory Coding (SPARC). SPARC comprises an articulatory analysis model that infers articulatory features from speech audio, and an articulatory synthesis model that synthesizes speech audio from articulatory features. The articulatory features are kinematic traces of vocal tract articulators and source features, which are intuitively interpretable and controllable, being the actual physical interface of speech production. An additional speaker identity encoder is jointly trained with the articulatory synthesizer to inform the voice texture of individual speakers. By training on large-scale speech data, we achieve a fully intelligible, high-quality articulatory synthesizer that generalizes to unseen speakers. Furthermore, the speaker embedding is effectively disentangled from articulations, which enables accent-perserving zero-shot voice conversion. To the best of our knowledge, this is the first demonstration of universal, high-performance articulatory inference and synthesis, suggesting the proposed framework as a powerful coding system of speech.
声道发音是语音产生的一个自然、基础的控制空间。发音器官的时空协调与声源相结合,形成了可理解的语音,从而实现有效的口语交流。基于语音的这一生理基础,我们提出了一种新的语音神经编码-解码框架--语音发音编码(SPARC)。SPARC 由一个发音分析模型和一个发音合成模型组成,前者可从语音音频中推断发音特征,后者可根据发音特征合成语音音频。发音特征是声道发音器的运动轨迹和声源特征,可直观解释和控制,是语音产生的实际物理界面。与发音合成器共同训练的还有一个说话者身份编码器,为单个说话者的声音质地提供信息。通过在大规模语音数据上进行训练,我们获得了一个完全可理解的高质量发音合成器,并可泛化到未见过的说话者。此外,说话人嵌入与发音有效分离,从而实现了口音保护零镜头语音转换。据我们所知,这是首次展示通用、高性能的发音推理和合成,表明所提出的框架是一种功能强大的语音编码系统。
{"title":"Coding Speech Through Vocal Tract Kinematics","authors":"Cheol Jun Cho;Peter Wu;Tejas S. Prabhune;Dhruv Agarwal;Gopala K. Anumanchipalli","doi":"10.1109/JSTSP.2024.3497655","DOIUrl":"https://doi.org/10.1109/JSTSP.2024.3497655","url":null,"abstract":"Vocal tract articulation is a natural, grounded control space of speech production. The spatiotemporal coordination of articulators combined with the vocal source shapes intelligible speech sounds to enable effective spoken communication. Based on this physiological grounding of speech, we propose a new framework of neural encoding-decoding of speech – Speech Articulatory Coding (SPARC). SPARC comprises an articulatory analysis model that infers articulatory features from speech audio, and an articulatory synthesis model that synthesizes speech audio from articulatory features. The articulatory features are kinematic traces of vocal tract articulators and source features, which are intuitively interpretable and controllable, being the actual physical interface of speech production. An additional speaker identity encoder is jointly trained with the articulatory synthesizer to inform the voice texture of individual speakers. By training on large-scale speech data, we achieve a fully intelligible, high-quality articulatory synthesizer that generalizes to unseen speakers. Furthermore, the speaker embedding is effectively disentangled from articulations, which enables accent-perserving zero-shot voice conversion. To the best of our knowledge, this is the first demonstration of universal, high-performance articulatory inference and synthesis, suggesting the proposed framework as a powerful coding system of speech.","PeriodicalId":13038,"journal":{"name":"IEEE Journal of Selected Topics in Signal Processing","volume":"18 8","pages":"1427-1440"},"PeriodicalIF":8.7,"publicationDate":"2024-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143184455","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DARIO: Differentiable Vision Transformer Pruning With Low-Cost Proxies DARIO:低成本代理的可微分视觉变压器修剪
IF 8.7 1区 工程技术 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-11-18 DOI: 10.1109/JSTSP.2024.3501685
Haozhe Sun;Alexandre Heuillet;Felix Mohr;Hedi Tabia
Transformer models have gained popularity for their exceptional performance. However, these models still face the challenge of high inference latency. To improve the computational efficiency of such models, we propose a novel differentiable pruning method called DARIO (DifferentiAble vision transformer pRunIng with low-cost prOxies). Our approach involves optimizing a set of gating parameters using differentiable, data-agnostic, scale-invariant, and low-cost performance proxies. DARIO is a data-agnostic pruning method, it does not need any classification heads during pruning. We evaluated DARIO on two popular state-of-the-art pre-trained ViT models, including both large (MAE-ViT) and small (MobileViT) sizes. Extensive experiments conducted across 40 diverse datasets demonstrated the effectiveness and efficiency of our DARIO method. DARIO not only significantly improves inference efficiency on modern hardware but also excels in preserving accuracy. Notably, DARIO has even achieved an increase in accuracy on MobileViT, despite only fine-tuning the last block and the classification head.
变压器型号因其卓越的性能而广受欢迎。然而,这些模型仍然面临着高推理延迟的挑战。为了提高这类模型的计算效率,我们提出了一种新的可微剪枝方法DARIO (differentiable vision transformer pruning with low-cost prOxies)。我们的方法包括使用可微的、数据不可知的、规模不变的和低成本的性能代理来优化一组门控参数。DARIO是一种与数据无关的剪枝方法,剪枝过程中不需要任何分类头。我们在两种流行的最先进的预训练ViT模型上评估了DARIO,包括大型(MAE-ViT)和小型(MobileViT)模型。在40个不同的数据集上进行的大量实验证明了我们的DARIO方法的有效性和效率。DARIO不仅显著提高了现代硬件上的推理效率,而且还保持了精度。值得注意的是,DARIO甚至在MobileViT上实现了准确性的提高,尽管只对最后一个块和分类头进行了微调。
{"title":"DARIO: Differentiable Vision Transformer Pruning With Low-Cost Proxies","authors":"Haozhe Sun;Alexandre Heuillet;Felix Mohr;Hedi Tabia","doi":"10.1109/JSTSP.2024.3501685","DOIUrl":"https://doi.org/10.1109/JSTSP.2024.3501685","url":null,"abstract":"Transformer models have gained popularity for their exceptional performance. However, these models still face the challenge of high inference latency. To improve the computational efficiency of such models, we propose a novel differentiable pruning method called DARIO (<bold>D</b>ifferenti<bold>A</b>ble vision transformer p<bold>R</b>un<bold>I</b>ng with low-cost pr<bold>O</b>xies). Our approach involves optimizing a set of gating parameters using differentiable, data-agnostic, scale-invariant, and low-cost performance proxies. DARIO is a data-agnostic pruning method, it does not need any classification heads during pruning. We evaluated DARIO on two popular state-of-the-art pre-trained ViT models, including both large (MAE-ViT) and small (MobileViT) sizes. Extensive experiments conducted across 40 diverse datasets demonstrated the effectiveness and efficiency of our DARIO method. DARIO not only significantly improves inference efficiency on modern hardware but also excels in preserving accuracy. Notably, DARIO has even achieved an increase in accuracy on MobileViT, despite only fine-tuning the last block and the classification head.","PeriodicalId":13038,"journal":{"name":"IEEE Journal of Selected Topics in Signal Processing","volume":"18 6","pages":"997-1009"},"PeriodicalIF":8.7,"publicationDate":"2024-11-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143106519","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improved Alias-and-Separate Speech Coding Framework With Minimal Algorithmic Delay 最小算法延迟的改进型别名和分离语音编码框架
IF 8.7 1区 工程技术 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-11-18 DOI: 10.1109/JSTSP.2024.3501681
Eunkyun Lee;Seungkwon Beack;Jong Won Shin
Alias-and-Separate (AaS) speech coding framework has shown the possibility to encode wideband (WB) speech with a narrowband (NB) speech codec and reconstruct it using speech separation. WB speech is first decimated incurring aliasing and then coded, transmitted, and decoded with a NB codec. The decoded signal is then separated into lower band and spectrally-flipped high band using a speech separation module, which are expanded, lowpass/highpass filtered, and added together to reconstruct the WB speech. The original AaS system, however, has algorithmic delay originated from the overlap-add operation for consecutive segments. This algorithmic delay can be reduced by omitting the overlap-add procedure, but the quality of the reconstructed speech is also degraded due to artifacts on the segment boundaries. In this work, we propose an improved AaS framework with minimum algorithmic delay. The decoded signal is first expanded by inserting zeros in-between samples before being processed by source separation module. As the expanded signal can be viewed as a summation of the frequency-shifted versions of the original signal, the decoded-and-expanded signal is then separated into the frequency-shifted signals, which are multiplied by complex exponentials and summed up to reconstruct the original signal. With carefully designed transposed convolution operation in the separation module, the proposed system requires minimal algorithmic delay while preventing discontinuity at the segment boundaries. Additionally, we propose to employ a generative vocoder to further improve the perceived quality and a modified multi-resolution short-time Fourier transform (MR-STFT) loss. Experimental results on the WB speech coding with a NB codec demonstrated that the proposed system outperformed the original AaS system and the existing WB speech codec in the subjective listening test. We have also shown that the proposed method can be applied when the decimation factor is not 2 in the experiment on the fullband speech coding with a WB codec.
别名分离(AaS)语音编码框架显示了用窄带语音编解码器对宽带语音进行编码并使用语音分离对其进行重构的可能性。WB语音首先被抽取产生混叠,然后用NB编解码器进行编码、传输和解码。然后使用语音分离模块将解码后的信号分离为低频段和频谱翻转的高频段,对其进行扩展、低通/高通滤波,并将其加在一起重建WB语音。然而,原始的AaS系统由于对连续段进行重叠添加操作而存在算法延迟。该算法可以通过省略重叠添加过程来减少延迟,但由于段边界上的伪影,重构语音的质量也会降低。在这项工作中,我们提出了一个具有最小算法延迟的改进的AaS框架。解码后的信号首先通过在采样之间插入零进行扩展,然后由源分离模块进行处理。由于扩展后的信号可以看作是原始信号的频移版本的总和,解码和扩展后的信号然后被分离成频移信号,这些频移信号乘以复指数并求和以重建原始信号。通过在分离模块中精心设计的转置卷积操作,所提出的系统需要最小的算法延迟,同时防止在段边界处的不连续。此外,我们建议采用生成式声码器来进一步提高感知质量和改进的多分辨率短时傅里叶变换(MR-STFT)损失。用NB编解码器进行WB语音编码的实验结果表明,该系统在主观听力测试中优于原有的AaS系统和现有的WB语音编解码器。在用WB编解码器进行全频带语音编码的实验中,我们也证明了该方法可以在抽取因子不为2的情况下应用。
{"title":"Improved Alias-and-Separate Speech Coding Framework With Minimal Algorithmic Delay","authors":"Eunkyun Lee;Seungkwon Beack;Jong Won Shin","doi":"10.1109/JSTSP.2024.3501681","DOIUrl":"https://doi.org/10.1109/JSTSP.2024.3501681","url":null,"abstract":"Alias-and-Separate (AaS) speech coding framework has shown the possibility to encode wideband (WB) speech with a narrowband (NB) speech codec and reconstruct it using speech separation. WB speech is first decimated incurring aliasing and then coded, transmitted, and decoded with a NB codec. The decoded signal is then separated into lower band and spectrally-flipped high band using a speech separation module, which are expanded, lowpass/highpass filtered, and added together to reconstruct the WB speech. The original AaS system, however, has algorithmic delay originated from the overlap-add operation for consecutive segments. This algorithmic delay can be reduced by omitting the overlap-add procedure, but the quality of the reconstructed speech is also degraded due to artifacts on the segment boundaries. In this work, we propose an improved AaS framework with minimum algorithmic delay. The decoded signal is first expanded by inserting zeros in-between samples before being processed by source separation module. As the expanded signal can be viewed as a summation of the frequency-shifted versions of the original signal, the decoded-and-expanded signal is then separated into the frequency-shifted signals, which are multiplied by complex exponentials and summed up to reconstruct the original signal. With carefully designed transposed convolution operation in the separation module, the proposed system requires minimal algorithmic delay while preventing discontinuity at the segment boundaries. Additionally, we propose to employ a generative vocoder to further improve the perceived quality and a modified multi-resolution short-time Fourier transform (MR-STFT) loss. Experimental results on the WB speech coding with a NB codec demonstrated that the proposed system outperformed the original AaS system and the existing WB speech codec in the subjective listening test. We have also shown that the proposed method can be applied when the decimation factor is not 2 in the experiment on the fullband speech coding with a WB codec.","PeriodicalId":13038,"journal":{"name":"IEEE Journal of Selected Topics in Signal Processing","volume":"18 8","pages":"1414-1426"},"PeriodicalIF":8.7,"publicationDate":"2024-11-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143184471","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cooperative Multi-Model Training for Personalized Federated Learning Over Heterogeneous Devices 异构设备上个性化联邦学习的多模型协作训练
IF 8.7 1区 工程技术 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-11-13 DOI: 10.1109/JSTSP.2024.3497660
Jian Xu;Shuo Wan;Yinchuan Li;Sichun Luo;Zhilin Chen;Yunfeng Shao;Zhitang Chen;Shao-Lun Huang;Linqi Song
Federated learning (FL) is an increasingly popular paradigm for protecting data privacy in machine learning systems. However, the data heterogeneity and high computation cost/latency are challenging barriers for employing FL in real-world applications with heterogeneous devices. In this paper, we propose a novel personalized FL framework named $mathtt {CompFL}$ allowing cooperative training of models with varied structures to mitigate those issues. First, $mathtt {CompFL}$ initializes a set of expert models in varied sizes and allows each client to choose one or multiple expert models for training according to their capacities. Second, $mathtt {CompFL}$ combines the model decoupling strategy and local-global feature alignment to mitigate the adverse impact of label heterogeneity, where clients only share the feature extractor part for each model architecture. Third, to encourage mutual enhancement of various models, knowledge distillation in local training is further applied to improve the overall performance. To make our framework workable in real systems, we implement it in both centralized settings with server-coordinated parallel training, and decentralized settings with newly developed device-to-device training-forwarding schemes. Extensive experiments on benchmark datasets are conducted to verify the potential of our framework for personalized FL over heterogeneous devices.
联邦学习(FL)是机器学习系统中保护数据隐私的一种日益流行的范例。然而,数据的异构性和高计算成本/延迟是在具有异构设备的实际应用中使用FL的挑战障碍。在本文中,我们提出了一个名为$mathtt {CompFL}$的新颖的个性化FL框架,允许具有不同结构的模型的协作训练来缓解这些问题。首先,$mathtt {CompFL}$初始化一组不同大小的专家模型,并允许每个客户端根据自己的能力选择一个或多个专家模型进行训练。其次,$mathtt {CompFL}$结合了模型解耦策略和局部-全局特征对齐,以减轻标签异质性的不利影响,其中客户端只共享每个模型架构的特征提取器部分。第三,为了鼓励各种模型的相互增强,进一步运用局部训练中的知识提炼来提高整体性能。为了使我们的框架在实际系统中可行,我们在集中式设置中实现了服务器协调的并行训练,并在分散设置中实现了新开发的设备到设备的训练转发方案。在基准数据集上进行了广泛的实验,以验证我们的框架在异构设备上个性化FL的潜力。
{"title":"Cooperative Multi-Model Training for Personalized Federated Learning Over Heterogeneous Devices","authors":"Jian Xu;Shuo Wan;Yinchuan Li;Sichun Luo;Zhilin Chen;Yunfeng Shao;Zhitang Chen;Shao-Lun Huang;Linqi Song","doi":"10.1109/JSTSP.2024.3497660","DOIUrl":"https://doi.org/10.1109/JSTSP.2024.3497660","url":null,"abstract":"Federated learning (FL) is an increasingly popular paradigm for protecting data privacy in machine learning systems. However, the data heterogeneity and high computation cost/latency are challenging barriers for employing FL in real-world applications with heterogeneous devices. In this paper, we propose a novel personalized FL framework named <inline-formula><tex-math>$mathtt {CompFL}$</tex-math></inline-formula> allowing cooperative training of models with varied structures to mitigate those issues. First, <inline-formula><tex-math>$mathtt {CompFL}$</tex-math></inline-formula> initializes a set of expert models in varied sizes and allows each client to choose one or multiple expert models for training according to their capacities. Second, <inline-formula><tex-math>$mathtt {CompFL}$</tex-math></inline-formula> combines the model decoupling strategy and local-global feature alignment to mitigate the adverse impact of label heterogeneity, where clients only share the feature extractor part for each model architecture. Third, to encourage mutual enhancement of various models, knowledge distillation in local training is further applied to improve the overall performance. To make our framework workable in real systems, we implement it in both centralized settings with server-coordinated parallel training, and decentralized settings with newly developed device-to-device training-forwarding schemes. Extensive experiments on benchmark datasets are conducted to verify the potential of our framework for personalized FL over heterogeneous devices.","PeriodicalId":13038,"journal":{"name":"IEEE Journal of Selected Topics in Signal Processing","volume":"19 1","pages":"195-207"},"PeriodicalIF":8.7,"publicationDate":"2024-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143512858","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Journal of Selected Topics in Signal Processing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1