首页 > 最新文献

Multimedia Systems最新文献

英文 中文
Accurate entropy modeling in learned image compression with joint enchanced SwinT and CNN 利用联合增强型 SwinT 和 CNN 在学习图像压缩中建立精确的熵模型
IF 3.9 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-07-07 DOI: 10.1007/s00530-024-01405-w
Dongjian Yang, Xiaopeng Fan, Xiandong Meng, Debin Zhao

Recently, learned image compression (LIC) has shown significant research potential. Most existing LIC methods are CNN-based or transformer-based or mixed. However, these LIC methods suffer from a certain degree of degradation in global attention performance, as CNN has limited-sized convolution kernels while window partitioning is applied to reduce computational complexity in transformer. This gives rise to the following two issues: (1) The main autoencoder (AE) and hyper AE exhibit limited transformation capabilities due to insufficient global modeling, making it challenging to improve the accuracy of coarse-grained entropy model. (2) The fine-grained entropy model struggles to adaptively utilize a larger range of contexts, because of weaker global modeling capability. In this paper, we propose the LIC with joint enhanced swin transformer (SwinT) and CNN to improve the entropy modeling accuracy. The key in the proposed method is that we enhance the global modeling ability of SwinT by introducing neighborhood window attention while maintaining an acceptable computational complexity and combines the local modeling ability of CNN to form the enhanced SwinT and CNN block (ESTCB). Specifically, we reconstruct the main AE and hyper AE of LIC based on ESTCB, enhancing their global transformation capabilities and resulting in a more accurate coarse-grained entropy model. Besides, we combine ESTCB with the checkerboard mask and the channel autoregressive model to develop a spatial then channel fine-grained entropy model, expanding the scope of LIC adaptive reference contexts. Comprehensive experiments demonstrate that our proposed method achieves state-of-the-art rate-distortion performance compared to existing LIC models.

最近,学习图像压缩(LIC)显示出巨大的研究潜力。现有的 LIC 方法大多基于 CNN 或变换器,或混合使用。然而,由于 CNN 的卷积核大小有限,而变换器则采用窗口分割来降低计算复杂度,因此这些 LIC 方法的全局注意力性能都有一定程度的下降。这就产生了以下两个问题:(1)由于全局建模不足,主自动编码器(AE)和超自动编码器(hyper AE)表现出有限的变换能力,这对提高粗粒度熵模型的精度带来了挑战。(2) 由于全局建模能力较弱,细粒度熵模型难以自适应地利用更大范围的上下文。本文提出了联合增强型swin transformer(SwinT)和 CNN 的 LIC,以提高熵模型的精度。该方法的关键在于,我们在保持可接受的计算复杂度的同时,通过引入邻域窗口注意增强了 SwinT 的全局建模能力,并结合了 CNN 的局部建模能力,形成了增强 SwinT 和 CNN 块(ESTCB)。具体来说,我们基于 ESTCB 重构了 LIC 的主 AE 和超 AE,增强了它们的全局变换能力,从而得到了更精确的粗粒度熵模型。此外,我们还将 ESTCB 与棋盘式掩码和信道自回归模型相结合,建立了空间信道细粒度熵模型,扩大了 LIC 自适应参考上下文的范围。综合实验证明,与现有的 LIC 模型相比,我们提出的方法实现了最先进的速率失真性能。
{"title":"Accurate entropy modeling in learned image compression with joint enchanced SwinT and CNN","authors":"Dongjian Yang, Xiaopeng Fan, Xiandong Meng, Debin Zhao","doi":"10.1007/s00530-024-01405-w","DOIUrl":"https://doi.org/10.1007/s00530-024-01405-w","url":null,"abstract":"<p>Recently, learned image compression (LIC) has shown significant research potential. Most existing LIC methods are CNN-based or transformer-based or mixed. However, these LIC methods suffer from a certain degree of degradation in global attention performance, as CNN has limited-sized convolution kernels while window partitioning is applied to reduce computational complexity in transformer. This gives rise to the following two issues: (1) The main autoencoder (AE) and hyper AE exhibit limited transformation capabilities due to insufficient global modeling, making it challenging to improve the accuracy of coarse-grained entropy model. (2) The fine-grained entropy model struggles to adaptively utilize a larger range of contexts, because of weaker global modeling capability. In this paper, we propose the LIC with joint enhanced swin transformer (SwinT) and CNN to improve the entropy modeling accuracy. The key in the proposed method is that we enhance the global modeling ability of SwinT by introducing neighborhood window attention while maintaining an acceptable computational complexity and combines the local modeling ability of CNN to form the enhanced SwinT and CNN block (ESTCB). Specifically, we reconstruct the main AE and hyper AE of LIC based on ESTCB, enhancing their global transformation capabilities and resulting in a more accurate coarse-grained entropy model. Besides, we combine ESTCB with the checkerboard mask and the channel autoregressive model to develop a spatial then channel fine-grained entropy model, expanding the scope of LIC adaptive reference contexts. Comprehensive experiments demonstrate that our proposed method achieves state-of-the-art rate-distortion performance compared to existing LIC models.</p>","PeriodicalId":51138,"journal":{"name":"Multimedia Systems","volume":"08 1","pages":""},"PeriodicalIF":3.9,"publicationDate":"2024-07-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141571163","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A multi-scale no-reference video quality assessment method based on transformer 基于变压器的多尺度无参考视频质量评估方法
IF 3.9 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-07-06 DOI: 10.1007/s00530-024-01403-y
Yingan Cui, Zonghua Yu, Yuqin Feng, Huaijun Wang, Junhuai Li

Video quality assessment is essential for optimizing user experience, enhancing network efficiency, supporting video production and editing, improving advertising effectiveness, and strengthening security in monitoring and other domains. Reacting to the prevailing focus of current research on video detail distortion while overlooking the temporal relationships between video frames and the impact of content-dependent characteristics of the human visual system on video quality, this paper proposes a multi-scale no-reference video quality assessment method based on transformer. On the one hand, spatial features of the video are extracted using a network that combines swin-transformer and deformable convolution, and further information preservation is achieved through mixed pooling of features in video frames. On the other hand, a pyramid aggregation module is utilized to merge long-term and short-term memories, enhancing the ability to capture temporal changes. Experimental results on public datasets such as KoNViD-1k, CVD2014, and LIVE-VQC demonstrate the effectiveness of the proposed method in video quality prediction.

视频质量评估对于优化用户体验、提高网络效率、支持视频制作和编辑、提高广告效果以及加强监控和其他领域的安全性至关重要。针对当前研究普遍关注视频细节失真,而忽视视频帧间的时间关系,以及人类视觉系统的内容依赖特性对视频质量的影响,本文提出了一种基于变换器的多尺度无参考视频质量评估方法。一方面,利用结合了swin-transformer和可变形卷积的网络提取视频的空间特征,并通过混合汇集视频帧中的特征进一步实现信息保存。另一方面,利用金字塔聚合模块合并长期记忆和短期记忆,增强捕捉时间变化的能力。在 KoNViD-1k、CVD2014 和 LIVE-VQC 等公开数据集上的实验结果表明了所提方法在视频质量预测中的有效性。
{"title":"A multi-scale no-reference video quality assessment method based on transformer","authors":"Yingan Cui, Zonghua Yu, Yuqin Feng, Huaijun Wang, Junhuai Li","doi":"10.1007/s00530-024-01403-y","DOIUrl":"https://doi.org/10.1007/s00530-024-01403-y","url":null,"abstract":"<p>Video quality assessment is essential for optimizing user experience, enhancing network efficiency, supporting video production and editing, improving advertising effectiveness, and strengthening security in monitoring and other domains. Reacting to the prevailing focus of current research on video detail distortion while overlooking the temporal relationships between video frames and the impact of content-dependent characteristics of the human visual system on video quality, this paper proposes a multi-scale no-reference video quality assessment method based on transformer. On the one hand, spatial features of the video are extracted using a network that combines swin-transformer and deformable convolution, and further information preservation is achieved through mixed pooling of features in video frames. On the other hand, a pyramid aggregation module is utilized to merge long-term and short-term memories, enhancing the ability to capture temporal changes. Experimental results on public datasets such as KoNViD-1k, CVD2014, and LIVE-VQC demonstrate the effectiveness of the proposed method in video quality prediction.</p>","PeriodicalId":51138,"journal":{"name":"Multimedia Systems","volume":"62 1","pages":""},"PeriodicalIF":3.9,"publicationDate":"2024-07-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141571167","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep learning based features extraction for facial gender classification using ensemble of machine learning technique 基于深度学习的特征提取,利用机器学习技术组合进行面部性别分类
IF 3.9 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-07-06 DOI: 10.1007/s00530-024-01399-5
Fazal Waris, Feipeng Da, Shanghuan Liu

Accurate and efficient gender recognition is an essential for many applications such as surveillance, security, and biometrics. Recently, deep learning techniques have made remarkable advancements in feature extraction and have become extensively implemented in various applications, including gender classification. However, despite the numerous studies conducted on the problem, correctly recognizing robust and essential features from face images and efficiently distinguishing them with high accuracy in the wild is still a challenging task for real-world applications. This article proposes an approach that combines deep learning and soft voting-based ensemble model to perform automatic gender classification with high accuracy in an unconstrained environment. In the proposed technique, a novel deep convolutional neural network (DCNN) was designed to extract 128 high-quality and accurate features from face images. The StandardScaler method was then used to pre-process these extracted features, and finally, these preprocessed features were classified with soft voting ensemble learning model combining the outputs from several machine learning classifiers such as random forest (RF), support vector machine (SVM), linear discriminant analysis (LDA), logistic regression (LR), gradient boosting classifier (GBC) and XGBoost to improve the prediction accuracy. The experimental study was performed on the UTK, label faces in the wild (LFW), Adience and FEI datasets. The results attained evidently show that the proposed approach outperforms all current approaches in terms of accuracy across all datasets.

准确、高效的性别识别对于监控、安全和生物识别等许多应用都至关重要。最近,深度学习技术在特征提取方面取得了显著进步,并广泛应用于各种应用中,包括性别分类。然而,尽管对这一问题进行了大量研究,但从人脸图像中正确识别稳健的基本特征,并在野外高精度地有效区分这些特征,对于现实世界的应用来说仍然是一项具有挑战性的任务。本文提出了一种将深度学习和基于软投票的集合模型相结合的方法,以在无约束环境中高精度地执行自动性别分类。在所提出的技术中,设计了一种新型深度卷积神经网络(DCNN),用于从人脸图像中提取 128 个高质量的准确特征。然后使用 StandardScaler 方法对这些提取的特征进行预处理,最后使用软投票集合学习模型对这些预处理的特征进行分类,该模型结合了多个机器学习分类器的输出,如随机森林(RF)、支持向量机(SVM)、线性判别分析(LDA)、逻辑回归(LR)、梯度提升分类器(GBC)和 XGBoost,以提高预测精度。实验研究是在UTK、野生人脸标签(LFW)、Adience和FEI数据集上进行的。实验结果明显表明,在所有数据集上,所提出的方法在准确性方面都优于所有现有方法。
{"title":"Deep learning based features extraction for facial gender classification using ensemble of machine learning technique","authors":"Fazal Waris, Feipeng Da, Shanghuan Liu","doi":"10.1007/s00530-024-01399-5","DOIUrl":"https://doi.org/10.1007/s00530-024-01399-5","url":null,"abstract":"<p>Accurate and efficient gender recognition is an essential for many applications such as surveillance, security, and biometrics. Recently, deep learning techniques have made remarkable advancements in feature extraction and have become extensively implemented in various applications, including gender classification. However, despite the numerous studies conducted on the problem, correctly recognizing robust and essential features from face images and efficiently distinguishing them with high accuracy in the wild is still a challenging task for real-world applications. This article proposes an approach that combines deep learning and soft voting-based ensemble model to perform automatic gender classification with high accuracy in an unconstrained environment. In the proposed technique, a novel deep convolutional neural network (DCNN) was designed to extract 128 high-quality and accurate features from face images. The StandardScaler method was then used to pre-process these extracted features, and finally, these preprocessed features were classified with soft voting ensemble learning model combining the outputs from several machine learning classifiers such as random forest (RF), support vector machine (SVM), linear discriminant analysis (LDA), logistic regression (LR), gradient boosting classifier (GBC) and XGBoost to improve the prediction accuracy. The experimental study was performed on the UTK, label faces in the wild (LFW), Adience and FEI datasets. The results attained evidently show that the proposed approach outperforms all current approaches in terms of accuracy across all datasets.</p>","PeriodicalId":51138,"journal":{"name":"Multimedia Systems","volume":"367 1","pages":""},"PeriodicalIF":3.9,"publicationDate":"2024-07-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141571166","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ViCLEVR: a visual reasoning dataset and hybrid multimodal fusion model for visual question answering in Vietnamese ViCLEVR:用于越南语视觉问题解答的视觉推理数据集和混合多模态融合模型
IF 3.9 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-07-06 DOI: 10.1007/s00530-024-01394-w
Khiem Vinh Tran, Hao Phu Phan, Kiet Van Nguyen, Ngan Luu Thuy Nguyen

In recent years, visual question answering (VQA) has gained significant attention for its diverse applications, including intelligent car assistance, aiding visually impaired individuals, and document image information retrieval using natural language queries. VQA requires effective integration of information from questions and images to generate accurate answers. Neural models for VQA have made remarkable progress on large-scale datasets, with a primary focus on resource-rich languages like English. To address this, we introduce the ViCLEVR dataset, a pioneering collection for evaluating various visual reasoning capabilities in Vietnamese while mitigating biases. The dataset comprises over 26,000 images and 30,000 question-answer pairs (QAs), each question annotated to specify the type of reasoning involved. Leveraging this dataset, we conduct a comprehensive analysis of contemporary visual reasoning systems, offering valuable insights into their strengths and limitations. Furthermore, we present PhoVIT, a comprehensive multimodal fusion that identifies objects in images based on questions. The architecture effectively employs transformers to enable simultaneous reasoning over textual and visual data, merging both modalities at an early model stage. The experimental findings demonstrate that our proposed model achieves state-of-the-art performance across four evaluation metrics.

近年来,视觉问题解答(VQA)因其多样化的应用而备受关注,其中包括智能汽车辅助、视障人士辅助以及使用自然语言查询的文档图像信息检索。VQA 需要有效整合问题和图像信息,以生成准确的答案。用于 VQA 的神经模型在大规模数据集上取得了显著进展,主要集中在英语等资源丰富的语言上。为了解决这个问题,我们引入了 ViCLEVR 数据集,这是一个用于评估越南语中各种视觉推理能力的开创性数据集,同时还能减少偏差。该数据集包含 26,000 多张图片和 30,000 个问题-答案对(QAs),每个问题都标注了具体的推理类型。利用该数据集,我们对当代视觉推理系统进行了全面分析,对其优势和局限性提出了宝贵的见解。此外,我们还介绍了 PhoVIT,这是一种全面的多模态融合系统,可根据问题识别图像中的物体。该架构有效地利用转换器实现了对文本和视觉数据的同步推理,在早期模型阶段就融合了两种模式。实验结果表明,我们提出的模型在四个评估指标上都达到了最先进的性能。
{"title":"ViCLEVR: a visual reasoning dataset and hybrid multimodal fusion model for visual question answering in Vietnamese","authors":"Khiem Vinh Tran, Hao Phu Phan, Kiet Van Nguyen, Ngan Luu Thuy Nguyen","doi":"10.1007/s00530-024-01394-w","DOIUrl":"https://doi.org/10.1007/s00530-024-01394-w","url":null,"abstract":"<p>In recent years, visual question answering (VQA) has gained significant attention for its diverse applications, including intelligent car assistance, aiding visually impaired individuals, and document image information retrieval using natural language queries. VQA requires effective integration of information from questions and images to generate accurate answers. Neural models for VQA have made remarkable progress on large-scale datasets, with a primary focus on resource-rich languages like English. To address this, we introduce the ViCLEVR dataset, a pioneering collection for evaluating various visual reasoning capabilities in Vietnamese while mitigating biases. The dataset comprises over 26,000 images and 30,000 question-answer pairs (QAs), each question annotated to specify the type of reasoning involved. Leveraging this dataset, we conduct a comprehensive analysis of contemporary visual reasoning systems, offering valuable insights into their strengths and limitations. Furthermore, we present PhoVIT, a comprehensive multimodal fusion that identifies objects in images based on questions. The architecture effectively employs transformers to enable simultaneous reasoning over textual and visual data, merging both modalities at an early model stage. The experimental findings demonstrate that our proposed model achieves state-of-the-art performance across four evaluation metrics.</p>","PeriodicalId":51138,"journal":{"name":"Multimedia Systems","volume":"73 1","pages":""},"PeriodicalIF":3.9,"publicationDate":"2024-07-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141571165","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Blind quality evaluator for multi-exposure fusion image via joint sparse features and complex-wavelet statistical characteristics 通过联合稀疏特征和复小波统计特征对多曝光融合图像进行盲质量评估
IF 3.9 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-07-05 DOI: 10.1007/s00530-024-01404-x
Benquan Yang, Yueli Cui, Lihong Liu, Guang Chen, Jiamin Xu, Junhao Lin

Multi-Exposure Fusion (MEF) technique aims to fuse multiple images taken from the same scene at different exposure levels into an image with more details. Although more and more MEF algorithms have been developed, how to effectively evaluate the quality of MEF images has not been thoroughly investigated. To address this issue, a blind quality evaluator for MEF image via joint sparse features and complex-wavelet statistical characteristics is developed. Specifically, considering that color and structure distortions are inevitably introduced during the MEF operations, we first train a color dictionary in the Lab color space based on the color perception mechanism of human visual system, and extract sparse perceptual features to capture the color and structure distortions. Given an MEF image to be evaluated, its components in both luminance and color channels are derived first. Subsequently, these obtained components are sparsely encoded using the trained color dictionary, and the perceived sparse features are extracted from the derived sparse coefficients. In addition, considering the insensitivity of sparse features towards weak structural information in images, complex steerable pyramid decomposition is further performed over the generated chromaticity map. Consequently, perceptual features of magnitude, phase and cross-scale structural similarity index are extracted from complex wavelet coefficients within the chromaticity map as quality-aware features. Experimental results demonstrate that our proposed metric outperforms the existing classic image quality evaluation metrics while maintaining high accordance with human visual perception.

多重曝光融合(MEF)技术旨在将同一场景中以不同曝光水平拍摄的多幅图像融合成一幅具有更多细节的图像。虽然已有越来越多的 MEF 算法被开发出来,但如何有效评估 MEF 图像的质量还没有得到深入研究。为了解决这个问题,我们开发了一种通过联合稀疏特征和复小波统计特征进行 MEF 图像质量盲评估的方法。具体来说,考虑到 MEF 操作过程中不可避免地会引入色彩和结构失真,我们首先根据人类视觉系统的色彩感知机制,在 Lab 色彩空间中训练色彩字典,并提取稀疏感知特征来捕捉色彩和结构失真。给定一幅待评估的 MEF 图像,首先得出其亮度和颜色通道的分量。随后,使用训练有素的色彩字典对这些获得的分量进行稀疏编码,并从获得的稀疏系数中提取感知稀疏特征。此外,考虑到稀疏特征对图像中的弱结构信息不敏感,还对生成的色度图进行了复杂的可转向金字塔分解。因此,从色度图中的复小波系数中提取了幅度、相位和跨尺度结构相似性指数等感知特征,作为质量感知特征。实验结果表明,我们提出的指标优于现有的经典图像质量评价指标,同时与人类视觉感知保持高度一致。
{"title":"Blind quality evaluator for multi-exposure fusion image via joint sparse features and complex-wavelet statistical characteristics","authors":"Benquan Yang, Yueli Cui, Lihong Liu, Guang Chen, Jiamin Xu, Junhao Lin","doi":"10.1007/s00530-024-01404-x","DOIUrl":"https://doi.org/10.1007/s00530-024-01404-x","url":null,"abstract":"<p>Multi-Exposure Fusion (MEF) technique aims to fuse multiple images taken from the same scene at different exposure levels into an image with more details. Although more and more MEF algorithms have been developed, how to effectively evaluate the quality of MEF images has not been thoroughly investigated. To address this issue, a blind quality evaluator for MEF image via joint sparse features and complex-wavelet statistical characteristics is developed. Specifically, considering that color and structure distortions are inevitably introduced during the MEF operations, we first train a color dictionary in the Lab color space based on the color perception mechanism of human visual system, and extract sparse perceptual features to capture the color and structure distortions. Given an MEF image to be evaluated, its components in both luminance and color channels are derived first. Subsequently, these obtained components are sparsely encoded using the trained color dictionary, and the perceived sparse features are extracted from the derived sparse coefficients. In addition, considering the insensitivity of sparse features towards weak structural information in images, complex steerable pyramid decomposition is further performed over the generated chromaticity map. Consequently, perceptual features of magnitude, phase and cross-scale structural similarity index are extracted from complex wavelet coefficients within the chromaticity map as quality-aware features. Experimental results demonstrate that our proposed metric outperforms the existing classic image quality evaluation metrics while maintaining high accordance with human visual perception.</p>","PeriodicalId":51138,"journal":{"name":"Multimedia Systems","volume":"86 1","pages":""},"PeriodicalIF":3.9,"publicationDate":"2024-07-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141551902","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Context-aware adaptive network for UDA semantic segmentation 用于 UDA 语义分割的语境感知自适应网络
IF 3.9 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-07-05 DOI: 10.1007/s00530-024-01397-7
Yu Yuan, Jinlong Shi, Xin Shu, Qiang Qian, Yunna Song, Zhen Ou, Dan Xu, Xin Zuo, YueCheng Yu, Yunhan Sun

Unsupervised Domain Adaptation (UDA) plays a pivotal role in enhancing the segmentation performance of models in the target domain by mitigating the domain shift between the source and target domains. However, Existing UDA image mix methods often overlook the contextual association between classes, limiting the segmentation capability of the model. To address this issue, we propose the context-aware adaptive network that enhances the model’s perception of contextual association information and maintains the contextual associations between different classes in mixed images, thereby improving the adaptability of the model. Firstly, we design a image mix strategy based on dynamic class correlation called DCCMix that constructs class correlation meta groups to preserve the contextual associations between different classes. Simultaneously, DCCMix dynamically adjusts the class proportion of the source domain within the mixed domain to gradually align with the distribution of the target domain, thereby improving training effectiveness. Secondly, the feature-wise fusion module and contextual feature-aware module are designed to better perceive contextual information of images and alleviate the issue of information loss during the feature extraction. Finally, we propose an adaptive class-edge weight to strengthen the segmentation ability of edge pixels in the model. Experimental results demonstrate that our proposed method achieves the mloU of 63.2% and 69.8% on two UDA benchmark tasks: SYNTHIA (rightarrow) Cityscapes and GTA (rightarrow) Cityscapes respectively. The code is available at https://github.com/yuheyuan/CAAN.

无监督领域适应(UDA)通过减轻源领域和目标领域之间的领域偏移,在提高模型在目标领域的分割性能方面发挥着关键作用。然而,现有的 UDA 图像混合方法往往忽略了类之间的上下文关联,从而限制了模型的分割能力。为解决这一问题,我们提出了上下文感知自适应网络,它能增强模型对上下文关联信息的感知,并保持混合图像中不同类别之间的上下文关联,从而提高模型的自适应能力。首先,我们设计了一种基于动态类别相关性的图像混合策略,称为 DCCMix,它可以构建类别相关性元组,以保留不同类别之间的上下文关联。同时,DCCMix 会动态调整混合域中源域的类别比例,使其逐渐与目标域的分布相一致,从而提高训练效果。其次,设计了特征融合模块和上下文特征感知模块,以更好地感知图像的上下文信息,缓解特征提取过程中的信息丢失问题。最后,我们提出了自适应类边缘权重,以加强模型中边缘像素的分割能力。实验结果表明,我们提出的方法在两个 UDA 基准任务上的 mloU 分别达到了 63.2% 和 69.8%:城市景观》和《GTA 城市景观》。代码见 https://github.com/yuheyuan/CAAN。
{"title":"Context-aware adaptive network for UDA semantic segmentation","authors":"Yu Yuan, Jinlong Shi, Xin Shu, Qiang Qian, Yunna Song, Zhen Ou, Dan Xu, Xin Zuo, YueCheng Yu, Yunhan Sun","doi":"10.1007/s00530-024-01397-7","DOIUrl":"https://doi.org/10.1007/s00530-024-01397-7","url":null,"abstract":"<p>Unsupervised Domain Adaptation (UDA) plays a pivotal role in enhancing the segmentation performance of models in the target domain by mitigating the domain shift between the source and target domains. However, Existing UDA image mix methods often overlook the contextual association between classes, limiting the segmentation capability of the model. To address this issue, we propose the context-aware adaptive network that enhances the model’s perception of contextual association information and maintains the contextual associations between different classes in mixed images, thereby improving the adaptability of the model. Firstly, we design a image mix strategy based on dynamic class correlation called DCCMix that constructs class correlation meta groups to preserve the contextual associations between different classes. Simultaneously, DCCMix dynamically adjusts the class proportion of the source domain within the mixed domain to gradually align with the distribution of the target domain, thereby improving training effectiveness. Secondly, the feature-wise fusion module and contextual feature-aware module are designed to better perceive contextual information of images and alleviate the issue of information loss during the feature extraction. Finally, we propose an adaptive class-edge weight to strengthen the segmentation ability of edge pixels in the model. Experimental results demonstrate that our proposed method achieves the mloU of 63.2% and 69.8% on two UDA benchmark tasks: SYNTHIA <span>(rightarrow)</span> Cityscapes and GTA <span>(rightarrow)</span> Cityscapes respectively. The code is available at https://github.com/yuheyuan/CAAN.</p>","PeriodicalId":51138,"journal":{"name":"Multimedia Systems","volume":"54 1","pages":""},"PeriodicalIF":3.9,"publicationDate":"2024-07-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141571164","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Quality evaluation methods of handwritten Chinese characters: a comprehensive survey 手写汉字质量评价方法:综合调查
IF 3.9 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-07-04 DOI: 10.1007/s00530-024-01396-8
Weiran Chen, Jiaqi Su, Weitao Song, Jialiang Xu, Guiqian Zhu, Ying Li, Yi Ji, Chunping Liu

Quality evaluation of handwritten Chinese characters aims to automatically quantify and assess handwritten Chinese characters through computer vision and machine learning technology. It is a topic of great concern for many handwriting learners and calligraphy enthusiasts. Over the past years, with the continuous development of computer technology, various new techniques have achieved flourishing and thriving progress. Nevertheless, how to realize fast and accurate character evaluation without human intervention is still one of the most challenging tasks in artificial intelligence. In this paper, we aim to provide a comprehensive survey of the existing handwritten Chinese character quality evaluation methods. Specifically, we first illustrate the research scope and background of the task. Then we outline our literature selection and analysis methodology, and review a series of related concepts, including common Chinese character features, evaluation metrics and classical machine learning models. After that, relying on the adopted mechanism and algorithm, we categorize the evaluation methods into two major groups: traditional methods and machine-learning-based methods. Representative approaches in each group are summarized, and their strengths and limitations are discussed in detail. Based on 191 papers in this survey, we finally conclude our paper with the challenges and future directions, with the expectation to provide some valuable illuminations for researchers in this field.

手写汉字质量评估旨在通过计算机视觉和机器学习技术自动量化和评估手写汉字。这是许多手写学习者和书法爱好者非常关注的话题。多年来,随着计算机技术的不断发展,各种新技术层出不穷、蓬勃发展。然而,如何在没有人工干预的情况下实现快速、准确的字符评估,仍然是人工智能领域最具挑战性的任务之一。本文旨在全面考察现有的手写汉字质量评价方法。具体来说,我们首先说明了这项任务的研究范围和背景。然后,我们概述了我们的文献选择和分析方法,并回顾了一系列相关概念,包括常见汉字特征、评价指标和经典机器学习模型。然后,根据所采用的机制和算法,我们将评价方法分为两大类:传统方法和基于机器学习的方法。我们总结了每一类中具有代表性的方法,并详细讨论了它们的优势和局限性。在 191 篇论文的基础上,我们最后总结了本文所面临的挑战和未来的发展方向,希望能为该领域的研究人员提供一些有价值的启示。
{"title":"Quality evaluation methods of handwritten Chinese characters: a comprehensive survey","authors":"Weiran Chen, Jiaqi Su, Weitao Song, Jialiang Xu, Guiqian Zhu, Ying Li, Yi Ji, Chunping Liu","doi":"10.1007/s00530-024-01396-8","DOIUrl":"https://doi.org/10.1007/s00530-024-01396-8","url":null,"abstract":"<p>Quality evaluation of handwritten Chinese characters aims to automatically quantify and assess handwritten Chinese characters through computer vision and machine learning technology. It is a topic of great concern for many handwriting learners and calligraphy enthusiasts. Over the past years, with the continuous development of computer technology, various new techniques have achieved flourishing and thriving progress. Nevertheless, how to realize fast and accurate character evaluation without human intervention is still one of the most challenging tasks in artificial intelligence. In this paper, we aim to provide a comprehensive survey of the existing handwritten Chinese character quality evaluation methods. Specifically, we first illustrate the research scope and background of the task. Then we outline our literature selection and analysis methodology, and review a series of related concepts, including common Chinese character features, evaluation metrics and classical machine learning models. After that, relying on the adopted mechanism and algorithm, we categorize the evaluation methods into two major groups: traditional methods and machine-learning-based methods. Representative approaches in each group are summarized, and their strengths and limitations are discussed in detail. Based on 191 papers in this survey, we finally conclude our paper with the challenges and future directions, with the expectation to provide some valuable illuminations for researchers in this field.</p>","PeriodicalId":51138,"journal":{"name":"Multimedia Systems","volume":"10 1","pages":""},"PeriodicalIF":3.9,"publicationDate":"2024-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141551799","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep contrastive multi-view clustering with doubly enhanced commonality 具有双重增强共性的深度对比多视角聚类
IF 3.9 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-07-04 DOI: 10.1007/s00530-024-01400-1
Zhiyuan Yang, Changming Zhu, Zishi Li

Recently, deep multi-view clustering leveraging autoencoders has garnered significant attention due to its ability to simultaneously enhance feature learning capabilities and optimize clustering outcomes. However, existing autoencoder-based deep multi-view clustering methods often exhibit a tendency to either overly emphasize view-specific information, thus neglecting shared information across views, or alternatively, to place undue focus on shared information, resulting in the dilution of complementary information from individual views. Given the principle that commonality resides within individuality, this paper proposes a staged training approach that comprises two phases: pre-training and fine-tuning. The pre-training phase primarily focuses on learning view-specific information, while the fine-tuning phase aims to doubly enhance commonality across views while maintaining these specific details. Specifically, we learn and extract the specific information of each view through the autoencoder in the pre-training stage. After entering the fine-tuning stage, we first initially enhance the commonality between independent specific views through the transformer layer, and then further strengthen these commonalities through contrastive learning on the semantic labels of each view, so as to obtain more accurate clustering results.

最近,利用自动编码器的深度多视图聚类方法因其能够同时增强特征学习能力和优化聚类结果而备受关注。然而,现有的基于自动编码器的深度多视图聚类方法往往表现出一种倾向,即过分强调视图的特定信息,从而忽略了视图间的共享信息;或者过分关注共享信息,从而稀释了单个视图的互补信息。鉴于共性寓于个性之中的原则,本文提出了一种分阶段训练方法,包括两个阶段:预训练和微调。预训练阶段主要侧重于学习特定视图的信息,而微调阶段的目的是在保持这些特定细节的同时,加倍增强不同视图之间的共性。具体来说,我们在预训练阶段通过自动编码器学习并提取每个视图的特定信息。进入微调阶段后,我们首先通过转换器层初步增强独立的特定视图之间的共性,然后通过对每个视图的语义标签进行对比学习来进一步强化这些共性,从而获得更准确的聚类结果。
{"title":"Deep contrastive multi-view clustering with doubly enhanced commonality","authors":"Zhiyuan Yang, Changming Zhu, Zishi Li","doi":"10.1007/s00530-024-01400-1","DOIUrl":"https://doi.org/10.1007/s00530-024-01400-1","url":null,"abstract":"<p>Recently, deep multi-view clustering leveraging autoencoders has garnered significant attention due to its ability to simultaneously enhance feature learning capabilities and optimize clustering outcomes. However, existing autoencoder-based deep multi-view clustering methods often exhibit a tendency to either overly emphasize view-specific information, thus neglecting shared information across views, or alternatively, to place undue focus on shared information, resulting in the dilution of complementary information from individual views. Given the principle that commonality resides within individuality, this paper proposes a staged training approach that comprises two phases: pre-training and fine-tuning. The pre-training phase primarily focuses on learning view-specific information, while the fine-tuning phase aims to doubly enhance commonality across views while maintaining these specific details. Specifically, we learn and extract the specific information of each view through the autoencoder in the pre-training stage. After entering the fine-tuning stage, we first initially enhance the commonality between independent specific views through the transformer layer, and then further strengthen these commonalities through contrastive learning on the semantic labels of each view, so as to obtain more accurate clustering results.</p>","PeriodicalId":51138,"journal":{"name":"Multimedia Systems","volume":"32 1","pages":""},"PeriodicalIF":3.9,"publicationDate":"2024-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141551798","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FEF-Net: feature enhanced fusion network with crossmodal attention for multimodal humor prediction FEF-Net:用于多模态幽默预测的跨模态注意力特征增强融合网络
IF 3.9 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-07-04 DOI: 10.1007/s00530-024-01402-z
Peng Gao, Chuanqi Tao, Donghai Guan

Humor segment prediction in video involves the comprehension and analysis of humor. Traditional humor prediction has been text-based; however, with the evolution of multimedia, the focus has shifted to multimodal approaches in humor prediction, marking a current trend in research. In recent years, determining whether a video is humorous has remained a challenge within the domain of sentiment analysis. Researchers have proposed multiple data fusion methods to address humor prediction and sentiment analysis. Within the realm of studying humor and emotions, text modality assumes a leading role, while audio and video modalities serve as supplementary data sources for multimodal humor prediction. However, these auxiliary modalities contain significant irrelevant information unrelated to the prediction task, resulting in redundancy. Current multimodal fusion models primarily emphasize fusion methods but overlook the issue of high redundancy in auxiliary modalities. The lack of research on reducing redundancy in auxiliary modalities introduces noise, thereby increasing the overall training complexity of models and diminishing predictive accuracy. Hence, developing a humor prediction method that effectively reduces redundancy in auxiliary modalities is pivotal for advancing multimodal research. In this paper, we propose the Feature Enhanced Fusion Network (FEF-Net), leveraging cross-modal attention to augment features from auxiliary modalities using knowledge from textual data. This mechanism generates weights to emphasize the redundancy of each corresponding time slice in the auxiliary modality. Further, employing Transformer encoders extracts high-level features for each modality, thereby enhancing the performance of humor prediction models. Experimental comparisons were conducted using the UR-FUNNY and MUStARD multimodal humor prediction models, revealing a 3.2% improvement in ‘Acc-2’ compared to the optimal model.

视频中的幽默片段预测涉及对幽默的理解和分析。传统的幽默预测以文本为基础,但随着多媒体的发展,幽默预测的重点已转向多模态方法,这标志着当前的研究趋势。近年来,判断视频是否幽默仍然是情感分析领域的一项挑战。研究人员提出了多种数据融合方法来解决幽默预测和情感分析问题。在幽默和情感研究领域,文本模式占据主导地位,而音频和视频模式则是多模式幽默预测的辅助数据源。然而,这些辅助模态包含大量与预测任务无关的信息,造成冗余。目前的多模态融合模型主要强调融合方法,却忽视了辅助模态的高冗余度问题。由于缺乏对减少辅助模态冗余的研究,从而引入了噪音,增加了模型的整体训练复杂度,降低了预测准确度。因此,开发一种能有效减少辅助模态冗余的幽默预测方法对于推进多模态研究至关重要。在本文中,我们提出了 "特征增强融合网络"(FEF-Net),利用跨模态注意力,利用文本数据的知识来增强辅助模态的特征。这种机制会生成权重,以强调辅助模态中每个相应时间片的冗余性。此外,采用 Transformer 编码器可提取每种模态的高级特征,从而提高幽默预测模型的性能。我们使用 UR-FUNNY 和 MUStARD 多模态幽默预测模型进行了实验比较,结果显示,与最优模型相比,"Acc-2 "提高了 3.2%。
{"title":"FEF-Net: feature enhanced fusion network with crossmodal attention for multimodal humor prediction","authors":"Peng Gao, Chuanqi Tao, Donghai Guan","doi":"10.1007/s00530-024-01402-z","DOIUrl":"https://doi.org/10.1007/s00530-024-01402-z","url":null,"abstract":"<p>Humor segment prediction in video involves the comprehension and analysis of humor. Traditional humor prediction has been text-based; however, with the evolution of multimedia, the focus has shifted to multimodal approaches in humor prediction, marking a current trend in research. In recent years, determining whether a video is humorous has remained a challenge within the domain of sentiment analysis. Researchers have proposed multiple data fusion methods to address humor prediction and sentiment analysis. Within the realm of studying humor and emotions, text modality assumes a leading role, while audio and video modalities serve as supplementary data sources for multimodal humor prediction. However, these auxiliary modalities contain significant irrelevant information unrelated to the prediction task, resulting in redundancy. Current multimodal fusion models primarily emphasize fusion methods but overlook the issue of high redundancy in auxiliary modalities. The lack of research on reducing redundancy in auxiliary modalities introduces noise, thereby increasing the overall training complexity of models and diminishing predictive accuracy. Hence, developing a humor prediction method that effectively reduces redundancy in auxiliary modalities is pivotal for advancing multimodal research. In this paper, we propose the Feature Enhanced Fusion Network (FEF-Net), leveraging cross-modal attention to augment features from auxiliary modalities using knowledge from textual data. This mechanism generates weights to emphasize the redundancy of each corresponding time slice in the auxiliary modality. Further, employing Transformer encoders extracts high-level features for each modality, thereby enhancing the performance of humor prediction models. Experimental comparisons were conducted using the UR-FUNNY and MUStARD multimodal humor prediction models, revealing a 3.2% improvement in ‘Acc-2’ compared to the optimal model.</p>","PeriodicalId":51138,"journal":{"name":"Multimedia Systems","volume":"44 1","pages":""},"PeriodicalIF":3.9,"publicationDate":"2024-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141551797","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A channel-gained single-model network with variable rate for multispectral image compression in UAV air-to-ground remote sensing 用于无人机空对地遥感多光谱图像压缩的具有可变速率的信道增益单模型网络
IF 3.9 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-07-02 DOI: 10.1007/s00530-024-01398-6
Wei Wang, Daiyin Zhu, Kedi Hu

Unmanned aerial vehicle (UAV) air-to-ground remote sensing technology, has the advantages of long flight duration, real-time image transmission, wide applicability, low cost, and so on. To better preserve the integrity of image features during transmission and storage, and improve efficiency in the meanwhile, image compression is a very important link. Nowadays the image compressor based on deep learning framework has been updating as the technological development. However, in order to obtain enough bit rates to fit the performance curve, there is always a severe computational burden, especially for multispectral image compression. This problem arises not only because the complexity of the algorithm is deepening, but also repeated training with rate-distortion optimization. In this paper, a channel-gained single-model network with variable rate for multispectral image compression is proposed. First, a channel gained module is introduced to map the channel content of the image to vector domain as amplitude factors, which leads to representation scaling, as well as obtaining the image representation of different bit rates in a single model. Second, after extracting spatial-spectral features, a plug-and-play dynamic response attention mechanism module is applied to take good care of distinguishing the content correlation of features and weighting the important area dynamically without adding extra parameters. Besides, a hyperprior autoencoder is used to make full use of edge information for entropy estimation, which contributes to a more accurate entropy model. The experiments prove that the proposed method greatly reduces the computational cost, while maintaining good compression performance and surpasses JPEG2000 and some other algorithms based on deep learning in PSNR, MSSSIM and MSA.

无人机(UAV)空对地遥感技术,具有飞行时间长、图像传输实时、适用性广、成本低等优点。为了在传输和存储过程中更好地保持图像特征的完整性,同时提高效率,图像压缩是一个非常重要的环节。如今,随着技术的发展,基于深度学习框架的图像压缩技术也在不断更新。然而,为了获得足够的比特率以适应性能曲线,始终存在着严重的计算负担,尤其是多光谱图像压缩。出现这一问题的原因不仅在于算法复杂度的不断加深,还在于反复训练的速率失真优化。本文提出了一种用于多光谱图像压缩的速率可变的信道增益单模型网络。首先,引入信道增益模块,将图像的信道内容以振幅因子的形式映射到矢量域,从而实现表示缩放,并在单一模型中获得不同比特率的图像表示。其次,在提取空间光谱特征后,应用即插即用的动态响应关注机制模块,在不增加额外参数的情况下,很好地区分特征的内容相关性,并对重要区域进行动态加权。此外,该方法还采用了超优先自动编码器,充分利用边缘信息进行熵估计,从而建立了更精确的熵模型。实验证明,所提出的方法大大降低了计算成本,同时保持了良好的压缩性能,在 PSNR、MSSSIM 和 MSA 方面超过了 JPEG2000 和其他一些基于深度学习的算法。
{"title":"A channel-gained single-model network with variable rate for multispectral image compression in UAV air-to-ground remote sensing","authors":"Wei Wang, Daiyin Zhu, Kedi Hu","doi":"10.1007/s00530-024-01398-6","DOIUrl":"https://doi.org/10.1007/s00530-024-01398-6","url":null,"abstract":"<p>Unmanned aerial vehicle (UAV) air-to-ground remote sensing technology, has the advantages of long flight duration, real-time image transmission, wide applicability, low cost, and so on. To better preserve the integrity of image features during transmission and storage, and improve efficiency in the meanwhile, image compression is a very important link. Nowadays the image compressor based on deep learning framework has been updating as the technological development. However, in order to obtain enough bit rates to fit the performance curve, there is always a severe computational burden, especially for multispectral image compression. This problem arises not only because the complexity of the algorithm is deepening, but also repeated training with rate-distortion optimization. In this paper, a channel-gained single-model network with variable rate for multispectral image compression is proposed. First, a channel gained module is introduced to map the channel content of the image to vector domain as amplitude factors, which leads to representation scaling, as well as obtaining the image representation of different bit rates in a single model. Second, after extracting spatial-spectral features, a plug-and-play dynamic response attention mechanism module is applied to take good care of distinguishing the content correlation of features and weighting the important area dynamically without adding extra parameters. Besides, a hyperprior autoencoder is used to make full use of edge information for entropy estimation, which contributes to a more accurate entropy model. The experiments prove that the proposed method greatly reduces the computational cost, while maintaining good compression performance and surpasses JPEG2000 and some other algorithms based on deep learning in PSNR, MSSSIM and MSA.</p>","PeriodicalId":51138,"journal":{"name":"Multimedia Systems","volume":"19 1","pages":""},"PeriodicalIF":3.9,"publicationDate":"2024-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141506765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Multimedia Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1