首页 > 最新文献

Journal of Electronic Imaging最新文献

英文 中文
End-to-end multitasking network for smart container product positioning and segmentation 用于智能集装箱产品定位和细分的端到端多任务网络
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-01 DOI: 10.1117/1.jei.33.5.053009
Wenzhong Shen, Xuejian Cai
The current smart cooler’s commodity identification system first locates the item being purchased, followed by feature extraction and matching. However, this method often suffers from inaccuracies due to the presence of background in the detection frame, leading to missed detections and misidentifications. To address these issues, we propose an end-to-end You Only Look Once (YOLO) for detection and segmentation algorithm. In the backbone network, we combine deformable convolution with a channel-to-pixel (C2f) module to enhance the model feature extraction capability. In the neck network, we introduce an optimized feature fusion structure, which is based on the weighted bi-directional feature pyramid. To further enhance the model’s understanding of both global and local context, a triple feature encoding module is employed, seamlessly fusing multi-scale features for improved performance. The convolutional block attention module is connected to the improved C2f module to enhance the network’s attention to the commodity image channel and spatial information. A supplementary segmentation branch is incorporated into the head of the network, allowing it to not only detect targets within the image but also generate precise segmentation masks for each detected object, thereby enhancing its multi-task capabilities. Compared with YOLOv8, for box and mask, the precision increases by 3% and 4.7%, recall increases by 2.8% and 4.7%, and mean average precision (mAP) increases by 4.9% and 14%. The frames per second is 119, which meets the demand for real-time detection. The results of comparative and ablation studies confirm the high accuracy and performance of the proposed algorithm, solidifying its foundation for fine-grained commodity identification.
目前智能冷柜的商品识别系统首先定位购买的商品,然后进行特征提取和匹配。然而,由于检测帧中存在背景,这种方法往往存在误差,从而导致漏检和错误识别。为了解决这些问题,我们提出了一种端到端的 "只看一遍"(YOLO)检测和分割算法。在骨干网络中,我们将可变形卷积与通道到像素(C2f)模块相结合,以增强模型特征提取能力。在颈部网络中,我们引入了基于加权双向特征金字塔的优化特征融合结构。为了进一步增强模型对全局和局部背景的理解,我们采用了三重特征编码模块,无缝融合多尺度特征以提高性能。卷积块关注模块与改进的 C2f 模块相连,以增强网络对商品图像通道和空间信息的关注。在网络的头部加入了一个辅助分割分支,使其不仅能检测图像中的目标,还能为每个检测到的物体生成精确的分割掩码,从而增强了其多任务处理能力。与 YOLOv8 相比,方框和掩码的精度分别提高了 3% 和 4.7%,召回率分别提高了 2.8% 和 4.7%,平均精度 (mAP) 分别提高了 4.9% 和 14%。每秒帧数为 119,满足了实时检测的要求。对比研究和消融研究的结果证实了所提算法的高精确度和高性能,为细粒度商品识别奠定了坚实的基础。
{"title":"End-to-end multitasking network for smart container product positioning and segmentation","authors":"Wenzhong Shen, Xuejian Cai","doi":"10.1117/1.jei.33.5.053009","DOIUrl":"https://doi.org/10.1117/1.jei.33.5.053009","url":null,"abstract":"The current smart cooler’s commodity identification system first locates the item being purchased, followed by feature extraction and matching. However, this method often suffers from inaccuracies due to the presence of background in the detection frame, leading to missed detections and misidentifications. To address these issues, we propose an end-to-end You Only Look Once (YOLO) for detection and segmentation algorithm. In the backbone network, we combine deformable convolution with a channel-to-pixel (C2f) module to enhance the model feature extraction capability. In the neck network, we introduce an optimized feature fusion structure, which is based on the weighted bi-directional feature pyramid. To further enhance the model’s understanding of both global and local context, a triple feature encoding module is employed, seamlessly fusing multi-scale features for improved performance. The convolutional block attention module is connected to the improved C2f module to enhance the network’s attention to the commodity image channel and spatial information. A supplementary segmentation branch is incorporated into the head of the network, allowing it to not only detect targets within the image but also generate precise segmentation masks for each detected object, thereby enhancing its multi-task capabilities. Compared with YOLOv8, for box and mask, the precision increases by 3% and 4.7%, recall increases by 2.8% and 4.7%, and mean average precision (mAP) increases by 4.9% and 14%. The frames per second is 119, which meets the demand for real-time detection. The results of comparative and ablation studies confirm the high accuracy and performance of the proposed algorithm, solidifying its foundation for fine-grained commodity identification.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142202392","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-head attention with reinforcement learning for supervised video summarization 多头注意力与强化学习用于有监督的视频总结
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-01 DOI: 10.1117/1.jei.33.5.053010
Bhakti Deepak Kadam, Ashwini Mangesh Deshpande
With the substantial surge in available internet video data, the intricate task of video summarization has consistently attracted the computer vision research community to summarize the videos meaningfully. Many recent summarization techniques leverage bidirectional long short-term memory for its proficiency in modeling temporal dependencies. However, its effectiveness is limited to short-duration video clips, typically up to 90 to 100 frames. To address this constraint, the proposed approach incorporates global and local multi-head attention, effectively capturing temporal dependencies at both global and local levels. This enhancement enables parallel computation, thereby improving overall performance for longer videos. This work considers video summarization as a supervised learning task and introduces a deep summarization architecture called multi-head attention with reinforcement learning (MHA-RL). The architecture comprises a pretrained convolutional neural network for extracting features from video frames, along with global and local multi-head attention mechanisms for predicting frame importance scores. Additionally, the network integrates an RL-based regressor network to consider the diversity and representativeness of the generated video summary. Extensive experimentation is conducted on benchmark datasets, such as TVSum and SumMe. The proposed method exhibits improved performance compared to the majority of state-of-the-art summarization techniques, as indicated by both qualitative and quantitative results.
随着可用互联网视频数据的激增,视频摘要这一复杂任务一直吸引着计算机视觉研究界对视频进行有意义的摘要。最近的许多摘要技术都利用了双向长时短时记忆在时间依赖性建模方面的优势。然而,其有效性仅限于短时视频片段,通常最多为 90 至 100 帧。为解决这一限制,所提出的方法结合了全局和局部多头注意力,有效捕捉了全局和局部层面的时间依赖性。这一改进实现了并行计算,从而提高了较长视频的整体性能。本研究将视频摘要视为一种监督学习任务,并引入了一种名为 "多头注意力与强化学习(MHA-RL)"的深度摘要架构。该架构包括一个用于从视频帧中提取特征的预训练卷积神经网络,以及用于预测帧重要性得分的全局和局部多头注意力机制。此外,该网络还集成了基于 RL 的回归网络,以考虑生成的视频摘要的多样性和代表性。在 TVSum 和 SumMe 等基准数据集上进行了广泛的实验。从定性和定量结果来看,与大多数最先进的摘要技术相比,所提出的方法表现出更高的性能。
{"title":"Multi-head attention with reinforcement learning for supervised video summarization","authors":"Bhakti Deepak Kadam, Ashwini Mangesh Deshpande","doi":"10.1117/1.jei.33.5.053010","DOIUrl":"https://doi.org/10.1117/1.jei.33.5.053010","url":null,"abstract":"With the substantial surge in available internet video data, the intricate task of video summarization has consistently attracted the computer vision research community to summarize the videos meaningfully. Many recent summarization techniques leverage bidirectional long short-term memory for its proficiency in modeling temporal dependencies. However, its effectiveness is limited to short-duration video clips, typically up to 90 to 100 frames. To address this constraint, the proposed approach incorporates global and local multi-head attention, effectively capturing temporal dependencies at both global and local levels. This enhancement enables parallel computation, thereby improving overall performance for longer videos. This work considers video summarization as a supervised learning task and introduces a deep summarization architecture called multi-head attention with reinforcement learning (MHA-RL). The architecture comprises a pretrained convolutional neural network for extracting features from video frames, along with global and local multi-head attention mechanisms for predicting frame importance scores. Additionally, the network integrates an RL-based regressor network to consider the diversity and representativeness of the generated video summary. Extensive experimentation is conducted on benchmark datasets, such as TVSum and SumMe. The proposed method exhibits improved performance compared to the majority of state-of-the-art summarization techniques, as indicated by both qualitative and quantitative results.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142202391","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient attention-based networks for fire and smoke detection 基于注意力的高效火灾和烟雾探测网络
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-01 DOI: 10.1117/1.jei.33.5.053014
Bowei Xiao, Chunman Yan
To address limitations in current flame and smoke detection models, including difficulties in handling irregularities, occlusions, large model sizes, and real-time performance issues, this work introduces FS-YOLO, a lightweight attention-based model. FS-YOLO adopts an efficient architecture for feature extraction capable of capturing long-range information, overcoming issues of redundant data and inadequate global feature extraction. The model incorporates squeeze-enhanced-axial-C2f to enhance global information capture without significantly increasing computational demands. Additionally, the improved VoVNet-GSConv-cross stage partial network refines semantic information from higher-level features, reducing missed detections and maintaining a lightweight model. Compared to YOLOv8n, FS-YOLO achieves a 1.4% increase and a 1.0% increase in mAP0.5 and mAP0.5:0.95, respectively, along with a 1.3% improvement in precision and a 1.0% boost in recall. These enhancements make FS-YOLO a promising solution for flame and smoke detection, balancing accuracy and efficiency effectively.
针对当前火焰和烟雾检测模型的局限性,包括难以处理不规则性、遮挡、模型尺寸过大和实时性等问题,本研究推出了基于注意力的轻量级模型 FS-YOLO。FS-YOLO 采用高效的特征提取架构,能够捕捉远距离信息,克服了冗余数据和全局特征提取不足的问题。该模型结合了挤压增强轴向-C2f,在不显著增加计算需求的情况下增强了全局信息捕捉能力。此外,改进后的 VoVNet-GSConv 跨阶段部分网络从更高层次的特征中提炼语义信息,减少了漏检并保持了轻量级模型。与 YOLOv8n 相比,FS-YOLO 的 mAP0.5 和 mAP0.5:0.95 分别提高了 1.4% 和 1.0%,精确度提高了 1.3%,召回率提高了 1.0%。这些改进使 FS-YOLO 成为一种很有前途的火焰和烟雾检测解决方案,有效地平衡了精度和效率。
{"title":"Efficient attention-based networks for fire and smoke detection","authors":"Bowei Xiao, Chunman Yan","doi":"10.1117/1.jei.33.5.053014","DOIUrl":"https://doi.org/10.1117/1.jei.33.5.053014","url":null,"abstract":"To address limitations in current flame and smoke detection models, including difficulties in handling irregularities, occlusions, large model sizes, and real-time performance issues, this work introduces FS-YOLO, a lightweight attention-based model. FS-YOLO adopts an efficient architecture for feature extraction capable of capturing long-range information, overcoming issues of redundant data and inadequate global feature extraction. The model incorporates squeeze-enhanced-axial-C2f to enhance global information capture without significantly increasing computational demands. Additionally, the improved VoVNet-GSConv-cross stage partial network refines semantic information from higher-level features, reducing missed detections and maintaining a lightweight model. Compared to YOLOv8n, FS-YOLO achieves a 1.4% increase and a 1.0% increase in mAP0.5 and mAP0.5:0.95, respectively, along with a 1.3% improvement in precision and a 1.0% boost in recall. These enhancements make FS-YOLO a promising solution for flame and smoke detection, balancing accuracy and efficiency effectively.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142202403","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DTSIDNet: a discrete wavelet and transformer based network for single image denoising DTSIDNet:基于离散小波和变换器的单幅图像去噪网络
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-01 DOI: 10.1117/1.jei.33.5.053007
Cong Hu, Yang Qu, Yuan-Bo Li, Xiao-Jun Wu
Recent advancements in transformer architectures have significantly enhanced image-denoising algorithms, surpassing the limitations of traditional convolutional neural networks by more effectively modeling global interactions through advanced attention mechanisms. In the domain of single-image denoising, noise manifests across various scales. This is especially evident in intricate scenarios, necessitating the comprehensive capture of multi-scale information inherent in the image. To solve transformer’s lack of multi-scale image analysis capability, a discrete wavelet and transformer based network (DTSIDNet) is proposed. The network adeptly resolves the inherent limitations of the transformer architecture by integrating the discrete wavelet transform. DTSIDNet independently manages image data at various scales, which greatly improves both adaptability and efficiency in environments with complex noise. The network’s self-attention mechanism dynamically shifts focus among different scales, efficiently capturing an extensive array of image features, thereby significantly enhancing the denoising outcome. This approach not only boosts the precision of denoising but also enhances the utilization of computational resources, striking an optimal balance between efficiency and high performance. Experiments on real-world and synthetic noise scenarios show that DTSIDNet delivers high image quality with low computational demands, indicating its superior performance in denoising tasks with efficient resource use.
变压器架构的最新进展大大增强了图像去噪算法,通过先进的注意力机制更有效地模拟全局交互,超越了传统卷积神经网络的局限性。在单图像去噪领域,噪声的表现形式多种多样。这在错综复杂的场景中尤为明显,需要全面捕捉图像中固有的多尺度信息。为了解决变压器缺乏多尺度图像分析能力的问题,我们提出了一种基于离散小波和变压器的网络(DTSIDNet)。该网络通过集成离散小波变换,巧妙地解决了变换器架构的固有局限性。DTSIDNet 可独立管理不同尺度的图像数据,从而大大提高了在复杂噪声环境中的适应性和效率。该网络的自我关注机制可在不同尺度之间动态转移焦点,有效捕捉大量图像特征,从而显著提高去噪效果。这种方法不仅提高了去噪的精度,还提高了计算资源的利用率,在效率和高性能之间取得了最佳平衡。在真实世界和合成噪声场景中的实验表明,DTSIDNet 能以较低的计算需求提供较高的图像质量,这表明它在去噪任务中具有高效利用资源的优越性能。
{"title":"DTSIDNet: a discrete wavelet and transformer based network for single image denoising","authors":"Cong Hu, Yang Qu, Yuan-Bo Li, Xiao-Jun Wu","doi":"10.1117/1.jei.33.5.053007","DOIUrl":"https://doi.org/10.1117/1.jei.33.5.053007","url":null,"abstract":"Recent advancements in transformer architectures have significantly enhanced image-denoising algorithms, surpassing the limitations of traditional convolutional neural networks by more effectively modeling global interactions through advanced attention mechanisms. In the domain of single-image denoising, noise manifests across various scales. This is especially evident in intricate scenarios, necessitating the comprehensive capture of multi-scale information inherent in the image. To solve transformer’s lack of multi-scale image analysis capability, a discrete wavelet and transformer based network (DTSIDNet) is proposed. The network adeptly resolves the inherent limitations of the transformer architecture by integrating the discrete wavelet transform. DTSIDNet independently manages image data at various scales, which greatly improves both adaptability and efficiency in environments with complex noise. The network’s self-attention mechanism dynamically shifts focus among different scales, efficiently capturing an extensive array of image features, thereby significantly enhancing the denoising outcome. This approach not only boosts the precision of denoising but also enhances the utilization of computational resources, striking an optimal balance between efficiency and high performance. Experiments on real-world and synthetic noise scenarios show that DTSIDNet delivers high image quality with low computational demands, indicating its superior performance in denoising tasks with efficient resource use.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142202390","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SMLoc: spatial multilayer perception-guided camera localization SMLoc:空间多层感知引导的相机定位
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-01 DOI: 10.1117/1.jei.33.5.053013
Jingyuan Feng, Shengsheng Wang, Haonan Sun
Camera localization is a technique for obtaining the camera’s six degrees of freedom using the camera as a sensor input. It is widely used in augmented reality, autonomous driving, virtual reality, etc. In recent years, with the development of deep-learning technology, absolute pose regression has gained wide attention as an end-to-end learning-based localization method. The typical architecture is constructed by a convolutional backbone and a multilayer perception (MLP) regression header composed of multiple fully connected layers. Typically, the two-dimensional feature maps extracted by the convolutional backbone have to be flattened and passed into the fully connected layer for pose regression. However, this operation will result in the loss of crucial pixel position information carried by the two-dimensional feature map and adversely affect the accuracy of the pose estimation. We propose a parallel structure, termed SMLoc, using a spatial MLP to aggregate position and orientation information from feature maps, respectively, reducing the loss of pixel position information. Our approach achieves superior performance on common indoor and outdoor datasets.
摄像头定位是一种利用摄像头作为传感器输入获取摄像头六自由度的技术。它被广泛应用于增强现实、自动驾驶、虚拟现实等领域。近年来,随着深度学习技术的发展,绝对姿态回归作为一种基于端到端学习的定位方法受到广泛关注。其典型架构由卷积骨干和多层全连接层组成的多层感知(MLP)回归头构建而成。通常情况下,卷积骨干层提取的二维特征图必须经过扁平化处理,然后传入全连接层进行姿态回归。然而,这种操作会导致二维特征图所携带的关键像素位置信息丢失,并对姿态估计的准确性产生不利影响。我们提出了一种名为 SMLoc 的并行结构,利用空间 MLP 分别聚合特征图中的位置和方向信息,从而减少像素位置信息的丢失。我们的方法在常见的室内和室外数据集上取得了优异的性能。
{"title":"SMLoc: spatial multilayer perception-guided camera localization","authors":"Jingyuan Feng, Shengsheng Wang, Haonan Sun","doi":"10.1117/1.jei.33.5.053013","DOIUrl":"https://doi.org/10.1117/1.jei.33.5.053013","url":null,"abstract":"Camera localization is a technique for obtaining the camera’s six degrees of freedom using the camera as a sensor input. It is widely used in augmented reality, autonomous driving, virtual reality, etc. In recent years, with the development of deep-learning technology, absolute pose regression has gained wide attention as an end-to-end learning-based localization method. The typical architecture is constructed by a convolutional backbone and a multilayer perception (MLP) regression header composed of multiple fully connected layers. Typically, the two-dimensional feature maps extracted by the convolutional backbone have to be flattened and passed into the fully connected layer for pose regression. However, this operation will result in the loss of crucial pixel position information carried by the two-dimensional feature map and adversely affect the accuracy of the pose estimation. We propose a parallel structure, termed SMLoc, using a spatial MLP to aggregate position and orientation information from feature maps, respectively, reducing the loss of pixel position information. Our approach achieves superior performance on common indoor and outdoor datasets.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142202401","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Toward effective local dimming-driven liquid crystal displays: a deep curve estimation–based adaptive compensation solution 实现有效的局部调光驱动液晶显示器:基于深度曲线估计的自适应补偿解决方案
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-01 DOI: 10.1117/1.jei.33.5.053005
Tianshan Liu, Kin-Man Lam
Local backlight dimming (LBD) is a promising technique for improving the contrast ratio and saving power consumption for liquid crystal displays. LBD consists of two crucial parts, i.e., backlight luminance determination, which locally controls the luminance of each sub-block of the backlight unit (BLU), and pixel compensation, which compensates for the reduction of pixel intensity, to achieve pleasing visual quality. However, the limitations of the current deep learning–based pixel compensation methods come from two aspects. First, it is difficult for a vanilla image-to-image translation strategy to learn the mapping relations between the input image and the compensated image, especially without considering the dimming levels. Second, the extensive model parameters make these methods hard to be deployed in industrial applications. To address these issues, we reformulate pixel compensation as an input-specific curve estimation task. Specifically, a deep lightweight network, namely, the curve estimation network (CENet), takes both the original input image and the dimmed BLUs as input, to estimate a set of high-order curves. Then, these curves are applied iteratively to adjust the intensity of each pixel to obtain a compensated image. Given the determined BLUs, the proposed CENet can be trained in an end-to-end manner. This implies that our proposed method is compatible with any backlight dimming strategies. Extensive evaluation results on the DIVerse 2K (DIV2K) dataset highlight the superiority of the proposed CENet-integrated local dimming framework, in terms of model size and visual quality of displayed content.
局部背光调光(LBD)是液晶显示器提高对比度和节省能耗的一项有前途的技术。局部背光调光由两个关键部分组成,即背光亮度确定和像素补偿,前者用于局部控制背光单元(BLU)每个子块的亮度,后者用于补偿像素强度的降低,以获得令人愉悦的视觉质量。然而,目前基于深度学习的像素补偿方法存在两方面的局限性。首先,虚构的图像到图像转换策略很难学习输入图像和补偿图像之间的映射关系,尤其是在不考虑调光等级的情况下。其次,大量的模型参数使得这些方法难以应用于工业领域。为了解决这些问题,我们将像素补偿重新表述为特定于输入的曲线估计任务。具体来说,一个深度轻量级网络,即曲线估算网络(CENet),将原始输入图像和调光 BLU 作为输入,估算出一组高阶曲线。然后,应用这些曲线迭代调整每个像素的强度,以获得补偿图像。有了确定的 BLU,建议的 CENet 就能以端到端的方式进行训练。这意味着我们提出的方法与任何背光调光策略都是兼容的。在 DIVerse 2K (DIV2K) 数据集上的广泛评估结果凸显了所提出的 CENet 集成局部调光框架在模型大小和显示内容的视觉质量方面的优越性。
{"title":"Toward effective local dimming-driven liquid crystal displays: a deep curve estimation–based adaptive compensation solution","authors":"Tianshan Liu, Kin-Man Lam","doi":"10.1117/1.jei.33.5.053005","DOIUrl":"https://doi.org/10.1117/1.jei.33.5.053005","url":null,"abstract":"Local backlight dimming (LBD) is a promising technique for improving the contrast ratio and saving power consumption for liquid crystal displays. LBD consists of two crucial parts, i.e., backlight luminance determination, which locally controls the luminance of each sub-block of the backlight unit (BLU), and pixel compensation, which compensates for the reduction of pixel intensity, to achieve pleasing visual quality. However, the limitations of the current deep learning–based pixel compensation methods come from two aspects. First, it is difficult for a vanilla image-to-image translation strategy to learn the mapping relations between the input image and the compensated image, especially without considering the dimming levels. Second, the extensive model parameters make these methods hard to be deployed in industrial applications. To address these issues, we reformulate pixel compensation as an input-specific curve estimation task. Specifically, a deep lightweight network, namely, the curve estimation network (CENet), takes both the original input image and the dimmed BLUs as input, to estimate a set of high-order curves. Then, these curves are applied iteratively to adjust the intensity of each pixel to obtain a compensated image. Given the determined BLUs, the proposed CENet can be trained in an end-to-end manner. This implies that our proposed method is compatible with any backlight dimming strategies. Extensive evaluation results on the DIVerse 2K (DIV2K) dataset highlight the superiority of the proposed CENet-integrated local dimming framework, in terms of model size and visual quality of displayed content.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142202394","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FCCText: frequency-color complementary bistream structure for scene text detection FCCText:用于场景文本检测的频色互补双稳态流结构
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-01 DOI: 10.1117/1.jei.33.4.043037
Ruiyi Han, Xin Li
Current scene text detection methods mainly employ RGB domain information for text localization, and their performance has not been fully exploited in many challenging scenes. Considering that the RGB features of text and background in complex environments are subtle and more discernible in the frequency domain, we consider that the frequency-domain information can effectively complement the RGB-domain features, collectively enhancing text detection capabilities. To this end, we propose a network with complementary frequency-domain semantic and color features, called the bistream structure, to facilitate text detection in scenes characterized by a wide variety of complex patterns. Our approach utilizes a frequency perception module (FPM) that converts features extracted by the backbone into the frequency domain to enhance the ability to distinguish the text from the complex background, thereby achieving coarse localization of texts. This innovation utilizes frequency-domain features to efficiently reveal text structures obscured by background noise in the RGB domain, resulting in a sharper differentiation between text and background elements in challenging scenarios. Moreover, we propose a complementary correction module that guides the fusion of multi-level RGB features through the coarse localization results, progressively refining the segmentation results to achieve the correction of the frequency domain features. Extensive experiments on the Total-Text, CTW1500, and MSRA-TD500 datasets demonstrate that our method achieves outstanding performance in scene text detection.
目前的场景文本检测方法主要采用 RGB 域信息进行文本定位,在许多具有挑战性的场景中,其性能尚未得到充分发挥。考虑到复杂环境中文字和背景的 RGB 特征比较微妙,在频域中更容易辨别,我们认为频域信息可以有效补充 RGB 域特征,共同提高文字检测能力。为此,我们提出了一种具有频域语义和颜色互补特征的网络,称为双流结构,以促进在具有各种复杂图案的场景中进行文本检测。我们的方法利用频率感知模块 (FPM),将骨干网提取的特征转换为频域特征,以增强从复杂背景中区分文本的能力,从而实现文本的粗略定位。这一创新利用频域特征来有效揭示被 RGB 域背景噪声掩盖的文本结构,从而在具有挑战性的场景中更清晰地区分文本和背景元素。此外,我们还提出了一个补充校正模块,通过粗略的定位结果引导多级 RGB 特征的融合,逐步完善分割结果,从而实现频域特征的校正。在 Total-Text、CTW1500 和 MSRA-TD500 数据集上进行的大量实验证明,我们的方法在场景文本检测中取得了出色的性能。
{"title":"FCCText: frequency-color complementary bistream structure for scene text detection","authors":"Ruiyi Han, Xin Li","doi":"10.1117/1.jei.33.4.043037","DOIUrl":"https://doi.org/10.1117/1.jei.33.4.043037","url":null,"abstract":"Current scene text detection methods mainly employ RGB domain information for text localization, and their performance has not been fully exploited in many challenging scenes. Considering that the RGB features of text and background in complex environments are subtle and more discernible in the frequency domain, we consider that the frequency-domain information can effectively complement the RGB-domain features, collectively enhancing text detection capabilities. To this end, we propose a network with complementary frequency-domain semantic and color features, called the bistream structure, to facilitate text detection in scenes characterized by a wide variety of complex patterns. Our approach utilizes a frequency perception module (FPM) that converts features extracted by the backbone into the frequency domain to enhance the ability to distinguish the text from the complex background, thereby achieving coarse localization of texts. This innovation utilizes frequency-domain features to efficiently reveal text structures obscured by background noise in the RGB domain, resulting in a sharper differentiation between text and background elements in challenging scenarios. Moreover, we propose a complementary correction module that guides the fusion of multi-level RGB features through the coarse localization results, progressively refining the segmentation results to achieve the correction of the frequency domain features. Extensive experiments on the Total-Text, CTW1500, and MSRA-TD500 datasets demonstrate that our method achieves outstanding performance in scene text detection.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141933511","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Chaotic multiple-image encryption scheme: a simple and highly efficient solution for diverse applications 混沌多图像加密方案:适用于各种应用的简单高效解决方案
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-01 DOI: 10.1117/1.jei.33.4.043032
K. Abhimanyu Kumar Patro, Pulkit Singh, Narendra Khatri, Bibhudendra Acharya
A multitude of confidential and personal digital images are commonly stored and transmitted by devices with limited resources. These devices necessitate the implementation of uncomplicated yet highly efficient encryption techniques to safeguard the images. The challenge of designing encryption algorithms for multiple digital images that are simple, secure, and highly efficient is significant. This challenge arises due to the large quantity of images involved and the considerable size and strong inter-pixel associations exhibited by these digital images. We propose a method for efficiently, simply, and securely encrypting multiple images simultaneously using chaotic one-dimensional (1D) maps. Initially, each grayscale image is consolidated into a single, substantial image. Through transpose columnar transposition and bit-XOR diffusion procedures, each block undergoes parallel permutation and diffusion. The incorporation of parallel permutation and diffusion functions accelerates and enhances the performance of the method. In contrast to existing multi-image encryption methods, the proposed approach consistently employs a single 1D chaotic map, rendering the algorithm both software and hardware efficient while maintaining simplicity. The encryption technique adheres to general requirements for simplicity and high efficiency. Security analysis and simulation results demonstrate that the proposed method is straightforward, highly efficient, and effectively enhances the security of cipher images.
大量机密和个人数字图像通常由资源有限的设备存储和传输。这些设备需要采用简单而高效的加密技术来保护图像。设计简单、安全、高效的多数字图像加密算法是一项重大挑战。由于涉及的图像数量庞大,而且这些数字图像的尺寸相当大,像素间的关联性也很强,因此出现了这一挑战。我们提出了一种利用混沌一维(1D)映射同时对多幅图像进行高效、简单和安全加密的方法。起初,每张灰度图像被合并成一张大图像。通过转置列转置和比特-XOR扩散程序,每个区块都会经历并行置换和扩散。并行置换和扩散函数的加入加速并提高了该方法的性能。与现有的多图像加密方法相比,所提出的方法始终采用单一的一维混沌图,使算法在保持简洁性的同时,还具有软件和硬件效率。该加密技术符合简单和高效的一般要求。安全分析和仿真结果表明,所提出的方法简单、高效,能有效提高密码图像的安全性。
{"title":"Chaotic multiple-image encryption scheme: a simple and highly efficient solution for diverse applications","authors":"K. Abhimanyu Kumar Patro, Pulkit Singh, Narendra Khatri, Bibhudendra Acharya","doi":"10.1117/1.jei.33.4.043032","DOIUrl":"https://doi.org/10.1117/1.jei.33.4.043032","url":null,"abstract":"A multitude of confidential and personal digital images are commonly stored and transmitted by devices with limited resources. These devices necessitate the implementation of uncomplicated yet highly efficient encryption techniques to safeguard the images. The challenge of designing encryption algorithms for multiple digital images that are simple, secure, and highly efficient is significant. This challenge arises due to the large quantity of images involved and the considerable size and strong inter-pixel associations exhibited by these digital images. We propose a method for efficiently, simply, and securely encrypting multiple images simultaneously using chaotic one-dimensional (1D) maps. Initially, each grayscale image is consolidated into a single, substantial image. Through transpose columnar transposition and bit-XOR diffusion procedures, each block undergoes parallel permutation and diffusion. The incorporation of parallel permutation and diffusion functions accelerates and enhances the performance of the method. In contrast to existing multi-image encryption methods, the proposed approach consistently employs a single 1D chaotic map, rendering the algorithm both software and hardware efficient while maintaining simplicity. The encryption technique adheres to general requirements for simplicity and high efficiency. Security analysis and simulation results demonstrate that the proposed method is straightforward, highly efficient, and effectively enhances the security of cipher images.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141885520","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Image-text multimodal classification via cross-attention contextual transformer with modality-collaborative learning 通过跨注意力上下文转换器与模态协作学习实现图像-文本多模态分类
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-01 DOI: 10.1117/1.jei.33.4.043042
Qianyao Shi, Wanru Xu, Zhenjiang Miao
Nowadays, we are surrounded by various types of data from different modalities, such as text, images, audio, and video. The existence of this multimodal data provides us with rich information, but it also brings new challenges: how do we effectively utilize this data for accurate classification? This is the main problem faced by multimodal classification tasks. Multimodal classification is an important task that aims to classify data from different modalities. However, due to the different characteristics and structures of data from different modalities, effectively fusing and utilizing them for classification is a challenging problem. To address this issue, we propose a cross-attention contextual transformer with modality-collaborative learning for multimodal classification (CACT-MCL-MMC) to better integrate information from different modalities. On the one hand, existing multimodal fusion methods ignore the intra- and inter-modality relationships, and there is unnoticed information in the modalities, resulting in unsatisfactory classification performance. To address the problem of insufficient interaction of modality information in existing algorithms, we use a cross-attention contextual transformer to capture the contextual relationships within and among modalities to improve the representativeness of the model. On the other hand, due to differences in the quality of information among different modalities, some modalities may have misleading or ambiguous information. Treating each modality equally may result in modality perceptual noise, which reduces the performance of multimodal classification. Therefore, we use modality-collaborative to filter misleading information, alleviate the quality difference of information among modalities, align modality information with high-quality and effective modalities, enhance unimodal information, and obtain more ideal multimodal fusion information to improve the model’s discriminative ability. Our comparative experimental results on two benchmark datasets for image-text classification, CrisisMMD and UPMC Food-101, show that our proposed model outperforms other classification methods and even state-of-the-art (SOTA) multimodal classification methods. Meanwhile, the effectiveness of the cross-attention module, multimodal contextual attention network, and modality-collaborative learning was verified through ablation experiments. In addition, conducting hyper-parameter validation experiments showed that different fusion calculation methods resulted in differences in experimental results. The most effective feature tensor calculation method was found. We also conducted qualitative experiments. Compared with the original model, our proposed model can identify the expected results in the vast majority of cases. The codes are available at https://github.com/KobeBryant8-24-MVP/CACT-MCL-MMC. The CrisisMMD is available at https://dataverse.mpisws.org/dataverse/icwsm18, and the UPMC-Food-101 is available at https://vi
如今,我们身边充斥着来自文本、图像、音频和视频等不同模式的各类数据。这些多模态数据的存在为我们提供了丰富的信息,但同时也带来了新的挑战:如何有效利用这些数据进行准确分类?这是多模态分类任务面临的主要问题。多模态分类是一项重要任务,旨在对来自不同模态的数据进行分类。然而,由于来自不同模态的数据具有不同的特征和结构,有效融合和利用这些数据进行分类是一个具有挑战性的问题。为了解决这个问题,我们提出了一种用于多模态分类的具有模态协作学习功能的跨注意力上下文转换器(CACT-MCL-MMC),以更好地整合来自不同模态的信息。一方面,现有的多模态融合方法忽视了模态内和模态间的关系,模态中存在未被关注的信息,导致分类效果不理想。针对现有算法中模态信息交互不足的问题,我们采用交叉注意上下文转换器来捕捉模态内和模态间的上下文关系,以提高模型的代表性。另一方面,由于不同模态之间的信息质量存在差异,某些模态的信息可能具有误导性或模糊性。对每种模态一视同仁可能会导致模态感知噪声,从而降低多模态分类的性能。因此,我们利用模态协同过滤误导信息,缓解模态间信息质量的差异,将模态信息与高质量、有效的模态信息相匹配,增强单模态信息,获得更理想的多模态融合信息,从而提高模型的判别能力。我们在 CrisisMMD 和 UPMC Food-101 这两个图像-文本分类基准数据集上的对比实验结果表明,我们提出的模型优于其他分类方法,甚至优于最先进的(SOTA)多模态分类方法。同时,交叉注意力模块、多模态情境注意力网络和模态协作学习的有效性也通过消融实验得到了验证。此外,超参数验证实验表明,不同的融合计算方法会导致实验结果的差异。我们找到了最有效的特征张量计算方法。我们还进行了定性实验。与原始模型相比,我们提出的模型在绝大多数情况下都能识别出预期结果。代码见 https://github.com/KobeBryant8-24-MVP/CACT-MCL-MMC。CrisisMMD 可在 https://dataverse.mpisws.org/dataverse/icwsm18 上查阅,UPMC-Food-101 可在 https://visiir.isir.upmc.fr/ 上查阅。
{"title":"Image-text multimodal classification via cross-attention contextual transformer with modality-collaborative learning","authors":"Qianyao Shi, Wanru Xu, Zhenjiang Miao","doi":"10.1117/1.jei.33.4.043042","DOIUrl":"https://doi.org/10.1117/1.jei.33.4.043042","url":null,"abstract":"Nowadays, we are surrounded by various types of data from different modalities, such as text, images, audio, and video. The existence of this multimodal data provides us with rich information, but it also brings new challenges: how do we effectively utilize this data for accurate classification? This is the main problem faced by multimodal classification tasks. Multimodal classification is an important task that aims to classify data from different modalities. However, due to the different characteristics and structures of data from different modalities, effectively fusing and utilizing them for classification is a challenging problem. To address this issue, we propose a cross-attention contextual transformer with modality-collaborative learning for multimodal classification (CACT-MCL-MMC) to better integrate information from different modalities. On the one hand, existing multimodal fusion methods ignore the intra- and inter-modality relationships, and there is unnoticed information in the modalities, resulting in unsatisfactory classification performance. To address the problem of insufficient interaction of modality information in existing algorithms, we use a cross-attention contextual transformer to capture the contextual relationships within and among modalities to improve the representativeness of the model. On the other hand, due to differences in the quality of information among different modalities, some modalities may have misleading or ambiguous information. Treating each modality equally may result in modality perceptual noise, which reduces the performance of multimodal classification. Therefore, we use modality-collaborative to filter misleading information, alleviate the quality difference of information among modalities, align modality information with high-quality and effective modalities, enhance unimodal information, and obtain more ideal multimodal fusion information to improve the model’s discriminative ability. Our comparative experimental results on two benchmark datasets for image-text classification, CrisisMMD and UPMC Food-101, show that our proposed model outperforms other classification methods and even state-of-the-art (SOTA) multimodal classification methods. Meanwhile, the effectiveness of the cross-attention module, multimodal contextual attention network, and modality-collaborative learning was verified through ablation experiments. In addition, conducting hyper-parameter validation experiments showed that different fusion calculation methods resulted in differences in experimental results. The most effective feature tensor calculation method was found. We also conducted qualitative experiments. Compared with the original model, our proposed model can identify the expected results in the vast majority of cases. The codes are available at https://github.com/KobeBryant8-24-MVP/CACT-MCL-MMC. The CrisisMMD is available at https://dataverse.mpisws.org/dataverse/icwsm18, and the UPMC-Food-101 is available at https://vi","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142202465","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SGTformer: improved Shifted Window Transformer network for white blood cell subtype classification SGTformer:用于白细胞亚型分类的改进型移位窗变换器网络
IF 1.1 4区 计算机科学 Q4 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-01 DOI: 10.1117/1.jei.33.4.043057
Xiangyu Deng, Lihao Pan, Zhiyan Dang
White blood cells are a core component of the immune system, responsible for protecting the human body from foreign invaders and infectious diseases. A decrease in the white blood cell count can lead to weakened immune function, increasing the risk of infection and illness. However, determining the number of white blood cells usually requires the expertise and effort of radiologists. In recent years, with the development of image processing technology, biomedical systems have widely applied image processing techniques in disease diagnosis. We aim to classify the subtypes of white blood cells using image processing technology. To improve the ability to extract fine information during the feature extraction process, the spatial prior convolutional attention (SPCA) module is proposed. In addition, to enhance the connection between features at distant distances, the Shifted Window (Swin) Transformer network is used as the backbone for feature extraction. The SGTformer network for white blood cell subtype classification is proposed by combining recursive gate convolution and SPCA modules. Our method is validated on the white blood cell dataset, and the experimental results demonstrate an overall accuracy of 99.47% in white blood cell classification, surpassing existing mainstream classification algorithms. It is evident that this method can effectively accomplish the task of white blood cell classification and provide robust support for the health of the immune system.
白细胞是免疫系统的核心组成部分,负责保护人体免受外来入侵者和传染病的侵害。白细胞数量减少会导致免疫功能减弱,增加感染和患病的风险。然而,确定白细胞数量通常需要放射科医生的专业知识和努力。近年来,随着图像处理技术的发展,生物医学系统已将图像处理技术广泛应用于疾病诊断。我们的目标是利用图像处理技术对白细胞进行亚型分类。为了提高特征提取过程中提取精细信息的能力,我们提出了空间先验卷积注意(SPCA)模块。此外,为了增强远距离特征之间的联系,还使用了移位窗(Swin)变换器网络作为特征提取的骨干。通过结合递归门卷积和 SPCA 模块,提出了用于白细胞亚型分类的 SGTformer 网络。我们的方法在白细胞数据集上进行了验证,实验结果表明白细胞分类的总体准确率为 99.47%,超过了现有的主流分类算法。由此可见,该方法能有效完成白细胞分类任务,为免疫系统的健康提供强有力的支持。
{"title":"SGTformer: improved Shifted Window Transformer network for white blood cell subtype classification","authors":"Xiangyu Deng, Lihao Pan, Zhiyan Dang","doi":"10.1117/1.jei.33.4.043057","DOIUrl":"https://doi.org/10.1117/1.jei.33.4.043057","url":null,"abstract":"White blood cells are a core component of the immune system, responsible for protecting the human body from foreign invaders and infectious diseases. A decrease in the white blood cell count can lead to weakened immune function, increasing the risk of infection and illness. However, determining the number of white blood cells usually requires the expertise and effort of radiologists. In recent years, with the development of image processing technology, biomedical systems have widely applied image processing techniques in disease diagnosis. We aim to classify the subtypes of white blood cells using image processing technology. To improve the ability to extract fine information during the feature extraction process, the spatial prior convolutional attention (SPCA) module is proposed. In addition, to enhance the connection between features at distant distances, the Shifted Window (Swin) Transformer network is used as the backbone for feature extraction. The SGTformer network for white blood cell subtype classification is proposed by combining recursive gate convolution and SPCA modules. Our method is validated on the white blood cell dataset, and the experimental results demonstrate an overall accuracy of 99.47% in white blood cell classification, surpassing existing mainstream classification algorithms. It is evident that this method can effectively accomplish the task of white blood cell classification and provide robust support for the health of the immune system.","PeriodicalId":54843,"journal":{"name":"Journal of Electronic Imaging","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142202438","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Electronic Imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1