首页 > 最新文献

IET Image Process.最新文献

英文 中文
An automatic feature selection and classification framework for analyzing ultrasound kidney images using dragonfly algorithm and random forest classifier 基于蜻蜓算法和随机森林分类器的肾脏超声图像自动特征选择与分类框架
Pub Date : 2021-03-22 DOI: 10.1049/IPR2.12179
C. Venkata Narasimhulu
In medical imaging, the automatic diagnosis of kidney carcinoma has become more diffi-cult because it is not easy to detect by physicians. Pre-processing is the first identification method to enhance image quality, remove noise and unwanted components from the back-drop of the kidneys image. The pre-processing method is essential and significant for the proposed algorithm. The objective of this analysis is to recognize and classify kidney dis-turbances with an ultrasound scan by providing a number of substantial content description parameters. The ultrasound pictures are prepared to protect the interest pixels before extracting the feature. A series of quantitative features were synthesized of each images, the principal component analysis was conducted for minimizing the number of features to produce set of wavelet-based multi-scale features. Dragonfly algorithm (DFA) was exe-cuted in this method. In the proposed work, the design and training of a random decision forest classifier and selected features are implemented. The classification of e-health information using ideal characteristics is used by the RF classifier. The proposed technique is activated in MATLAB/simulink work site and the experimental results show that the peak accuracy of the proposed technique is 95.6% using GWO-FFBN techniques compared to other existing techniques.
在医学影像学中,肾癌的自动诊断变得越来越困难,因为它不容易被医生发现。预处理是提高图像质量,去除肾脏图像背景噪声和不需要成分的第一步识别方法。预处理方法对该算法至关重要。本分析的目的是通过提供一些实质性的内容描述参数,通过超声扫描识别和分类肾脏紊乱。在提取特征之前,对超声图像进行预处理以保护感兴趣的像素点。对每张图像合成一系列定量特征,进行主成分分析,使特征数量最小化,得到一组基于小波的多尺度特征。该方法执行蜻蜓算法(DFA)。在本文提出的工作中,实现了随机决策森林分类器的设计和训练,并选择了特征。射频分类器使用理想特征对电子卫生信息进行分类。在MATLAB/simulink工作现场对该技术进行了激活,实验结果表明,与其他现有技术相比,使用GWO-FFBN技术所提出的技术的峰值精度为95.6%。
{"title":"An automatic feature selection and classification framework for analyzing ultrasound kidney images using dragonfly algorithm and random forest classifier","authors":"C. Venkata Narasimhulu","doi":"10.1049/IPR2.12179","DOIUrl":"https://doi.org/10.1049/IPR2.12179","url":null,"abstract":"In medical imaging, the automatic diagnosis of kidney carcinoma has become more diffi-cult because it is not easy to detect by physicians. Pre-processing is the first identification method to enhance image quality, remove noise and unwanted components from the back-drop of the kidneys image. The pre-processing method is essential and significant for the proposed algorithm. The objective of this analysis is to recognize and classify kidney dis-turbances with an ultrasound scan by providing a number of substantial content description parameters. The ultrasound pictures are prepared to protect the interest pixels before extracting the feature. A series of quantitative features were synthesized of each images, the principal component analysis was conducted for minimizing the number of features to produce set of wavelet-based multi-scale features. Dragonfly algorithm (DFA) was exe-cuted in this method. In the proposed work, the design and training of a random decision forest classifier and selected features are implemented. The classification of e-health information using ideal characteristics is used by the RF classifier. The proposed technique is activated in MATLAB/simulink work site and the experimental results show that the peak accuracy of the proposed technique is 95.6% using GWO-FFBN techniques compared to other existing techniques.","PeriodicalId":13486,"journal":{"name":"IET Image Process.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79317372","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
A robust sperm cell tracking algorithm using uneven lighting image fixing and improved branch and bound algorithm 基于不均匀光照图像固定和改进分支定界算法的精子细胞鲁棒跟踪算法
Pub Date : 2021-03-19 DOI: 10.1049/IPR2.12178
Ahmad Alhaj Alabdulla, A. Hasiloglu, E. Aksu
An accurate and robust sperm cells tracking algorithm that is able to detect and track sperm cells in videos with high accuracy and efficiency is presented. It is fast enough to process approximately 30 frames per second. It can find the correct path and measure motility parameters for each sperm. It can also adapt with different types of images coming from different cameras and bad recording conditions. Specifically, a new way is offered to optimize uneven lighting images to improve sperm cells detection which gives us the ability to get more accurate tracking results. The shape of each detected object is used to specify collided sperms and utilized dynamic gates which become bigger and smaller according to the sperm cell’s speed. For assigning tracks to the detected sperm cells positions an improved version of branch and bound algorithm which is faster than the normal one is offered. This sperm cells tracking algorithm outperforms many of the previous algorithms as it has lower error rate in both sperm detection and tracking. It is compared with six other algorithms, and it gives lower tracking error rates. This method will allow doctors and researchers to obtain sperm motility data instantly and accurately.
提出了一种精确、鲁棒的精子细胞跟踪算法,能够高精度、高效率地检测和跟踪视频中的精子细胞。它足够快,每秒可以处理大约30帧。它可以找到正确的路径并测量每个精子的运动参数。它还可以适应来自不同相机的不同类型的图像和恶劣的记录条件。具体来说,提供了一种新的方法来优化不均匀光照图像,以提高精子细胞的检测,使我们能够获得更准确的跟踪结果。每个被检测物体的形状被用来指定碰撞的精子,并利用动态门根据精子的速度变大或变小。为了给检测到的精子定位分配轨迹,提出了一种改进的分支绑定算法,该算法比常规算法更快。该算法在精子检测和跟踪方面具有较低的错误率,优于以往的许多算法。与其他六种算法进行了比较,结果表明该算法具有较低的跟踪错误率。这种方法将使医生和研究人员能够即时准确地获得精子运动数据。
{"title":"A robust sperm cell tracking algorithm using uneven lighting image fixing and improved branch and bound algorithm","authors":"Ahmad Alhaj Alabdulla, A. Hasiloglu, E. Aksu","doi":"10.1049/IPR2.12178","DOIUrl":"https://doi.org/10.1049/IPR2.12178","url":null,"abstract":"An accurate and robust sperm cells tracking algorithm that is able to detect and track sperm cells in videos with high accuracy and efficiency is presented. It is fast enough to process approximately 30 frames per second. It can find the correct path and measure motility parameters for each sperm. It can also adapt with different types of images coming from different cameras and bad recording conditions. Specifically, a new way is offered to optimize uneven lighting images to improve sperm cells detection which gives us the ability to get more accurate tracking results. The shape of each detected object is used to specify collided sperms and utilized dynamic gates which become bigger and smaller according to the sperm cell’s speed. For assigning tracks to the detected sperm cells positions an improved version of branch and bound algorithm which is faster than the normal one is offered. This sperm cells tracking algorithm outperforms many of the previous algorithms as it has lower error rate in both sperm detection and tracking. It is compared with six other algorithms, and it gives lower tracking error rates. This method will allow doctors and researchers to obtain sperm motility data instantly and accurately.","PeriodicalId":13486,"journal":{"name":"IET Image Process.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-03-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86183121","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
An optimized YOLO-based object detection model for crop harvesting system 一种优化的基于yolo的作物收获目标检测模型
Pub Date : 2021-03-18 DOI: 10.1049/IPR2.12181
M. H. Junos, A. S. M. Khairuddin, Subbiah Thannirmalai, M. Dahari
Funding information RU Grant-Faculty Programme by Faculty of Engineering, University of Malaya, Grant/Award Number: GPF042A-2019; Industry-Driven Innovation, Grant/Award Number: (IDIG)-PPSI-2020CLUSTER-SD01 Abstract The adoption of automated crop harvesting system based on machine vision may improve productivity and optimize the operational cost. The scope of this study is to obtain visual information at the plantation which is crucial in developing an intelligent automated crop harvesting system. This paper aims to develop an automatic detection system with high accuracy performance, low computational cost and lightweight model. Considering the advantages of YOLOv3 tiny, an optimized YOLOv3 tiny network namely YOLO-P is proposed to detect and localize three objects at palm oil plantation which include fresh fruit bunch, grabber and palm tree under various environment conditions. The proposed YOLO-P model incorporated lightweight backbone based on densely connected neural network, multi-scale detection architecture and optimized anchor box size. The experimental results demonstrated that the proposed YOLO-P model achieved good mean average precision and F1 score of 98.68% and 0.97 respectively. Besides, the proposed model performed faster training process and generated lightweight model of 76 MB. The proposed model was also tested to identify fresh fruit bunch of various maturities with accuracy of 98.91%. The comprehensive experimental results show that the proposed YOLO-P model can effectively perform robust and accurate detection at the palm oil plantation.
马来亚大学工程学院RU资助计划,资助/奖励编号:GPF042A-2019;摘要采用基于机器视觉的自动化作物收获系统可以提高生产力,优化运营成本。本研究的范围是在种植园内获取视觉信息,这对于开发智能自动化作物收获系统至关重要。本文旨在开发一种精度高、计算成本低、模型轻量化的自动检测系统。考虑到YOLOv3微型网络的优点,提出了一种优化的YOLOv3微型网络YOLO-P,用于在不同环境条件下对棕榈油种植园的新鲜果串、抓取器和棕榈树三种目标进行检测和定位。提出的YOLO-P模型结合了基于密集连接神经网络的轻型骨干、多尺度检测架构和优化锚盒尺寸。实验结果表明,所提出的YOLO-P模型取得了良好的平均精度,F1得分分别为98.68%和0.97。此外,该模型的训练速度更快,生成了76 MB的轻量级模型。对该模型进行了识别不同成熟度的新鲜水果串的测试,准确率达到98.91%。综合实验结果表明,所提出的YOLO-P模型能够有效地实现棕榈油种植园的鲁棒性和准确性检测。
{"title":"An optimized YOLO-based object detection model for crop harvesting system","authors":"M. H. Junos, A. S. M. Khairuddin, Subbiah Thannirmalai, M. Dahari","doi":"10.1049/IPR2.12181","DOIUrl":"https://doi.org/10.1049/IPR2.12181","url":null,"abstract":"Funding information RU Grant-Faculty Programme by Faculty of Engineering, University of Malaya, Grant/Award Number: GPF042A-2019; Industry-Driven Innovation, Grant/Award Number: (IDIG)-PPSI-2020CLUSTER-SD01 Abstract The adoption of automated crop harvesting system based on machine vision may improve productivity and optimize the operational cost. The scope of this study is to obtain visual information at the plantation which is crucial in developing an intelligent automated crop harvesting system. This paper aims to develop an automatic detection system with high accuracy performance, low computational cost and lightweight model. Considering the advantages of YOLOv3 tiny, an optimized YOLOv3 tiny network namely YOLO-P is proposed to detect and localize three objects at palm oil plantation which include fresh fruit bunch, grabber and palm tree under various environment conditions. The proposed YOLO-P model incorporated lightweight backbone based on densely connected neural network, multi-scale detection architecture and optimized anchor box size. The experimental results demonstrated that the proposed YOLO-P model achieved good mean average precision and F1 score of 98.68% and 0.97 respectively. Besides, the proposed model performed faster training process and generated lightweight model of 76 MB. The proposed model was also tested to identify fresh fruit bunch of various maturities with accuracy of 98.91%. The comprehensive experimental results show that the proposed YOLO-P model can effectively perform robust and accurate detection at the palm oil plantation.","PeriodicalId":13486,"journal":{"name":"IET Image Process.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80785606","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
An image encryption algorithm with a plaintext-related quantisation scheme 一个图像加密算法与明文相关的量化方案
Pub Date : 2021-03-18 DOI: 10.1049/IPR2.12174
Jakub Oravec, Ľ. Ovseník, J. Papaj
This paper describes an image encryption algorithm that utilises a plaintext-related quantisation scheme. Various plaintext-related approaches from other algorithms are presented and their properties are briefly discussed. Main advantage of the proposed solution is the achievement of a similar behaviour like that of more complex approaches with a plaintext-related technique used in a rather simple step such as quantisation. This design should result in a favourable computational complexity of the whole algorithm. The properties of the proposal are evaluated by a number of commonly used numerical parameters. Also, the statistical properties of a pseudo-random sequence that is quantised according to the plain image pixel intensities are investigated by tests from NIST 800-22 suite. Obtained results are compared to values reported in related works and they imply that the proposed solution produces encrypted images with comparable statistical properties but authors’ design is faster and more efficient.
本文描述了一种利用明文相关量化方案的图像加密算法。从其他算法中提出了各种与明文相关的方法,并简要讨论了它们的性质。所提出的解决方案的主要优点是,通过在相当简单的步骤(如量化)中使用与明文相关的技术,实现了与更复杂的方法类似的行为。这种设计将使整个算法具有良好的计算复杂度。该提案的性质是由一些常用的数值参数进行评估。此外,通过NIST 800-22套件的测试,研究了根据平面图像像素强度量化的伪随机序列的统计特性。将得到的结果与相关工作报告的值进行比较,结果表明所提出的解决方案产生的加密图像具有可比的统计特性,但作者的设计更快,更有效。
{"title":"An image encryption algorithm with a plaintext-related quantisation scheme","authors":"Jakub Oravec, Ľ. Ovseník, J. Papaj","doi":"10.1049/IPR2.12174","DOIUrl":"https://doi.org/10.1049/IPR2.12174","url":null,"abstract":"This paper describes an image encryption algorithm that utilises a plaintext-related quantisation scheme. Various plaintext-related approaches from other algorithms are presented and their properties are briefly discussed. Main advantage of the proposed solution is the achievement of a similar behaviour like that of more complex approaches with a plaintext-related technique used in a rather simple step such as quantisation. This design should result in a favourable computational complexity of the whole algorithm. The properties of the proposal are evaluated by a number of commonly used numerical parameters. Also, the statistical properties of a pseudo-random sequence that is quantised according to the plain image pixel intensities are investigated by tests from NIST 800-22 suite. Obtained results are compared to values reported in related works and they imply that the proposed solution produces encrypted images with comparable statistical properties but authors’ design is faster and more efficient.","PeriodicalId":13486,"journal":{"name":"IET Image Process.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73384948","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Scale space Radon transform 尺度空间Radon变换
Pub Date : 2021-03-18 DOI: 10.1049/IPR2.12180
D. Ziou, Nafaa Nacereddine, A. Goumeidane
An extension of Radon transform by using a measure function capturing the user need is proposed. The new transform, called scale space Radon transform, is devoted to the case where the embedded shape in the image is not filiform. A case study is brought on a straight line and an ellipse where the SSRT behaviour in the scale space and in the presence of noise is deeply analyzed. In order to show the effectiveness of the proposed transform, the experiments have been carried out, first, on linear and elliptical structures generated synthetically subjected to strong altering conditions such blur and noise and then on structures images issued from real-world applications such as road traffic, satellite imagery and weld X-ray imaging. Comparisons in terms of detection accuracy and computational time with well-known transforms and recent work dedicated to this purpose are conducted, where the proposed transform shows an outstanding performance in detecting the above-mentioned structures and targeting accurately their spatial locations even in low-quality images.
提出了利用捕获用户需求的度量函数对Radon变换进行扩展。这种新的变换称为尺度空间Radon变换,专门用于图像中嵌入的形状不是丝状的情况。以直线和椭圆为例,深入分析了在尺度空间和噪声存在下的SSRT行为。为了证明所提出的变换的有效性,首先对在模糊和噪声等强变化条件下合成的线性和椭圆结构进行了实验,然后对来自实际应用的结构图像进行了实验,如道路交通、卫星图像和焊缝x射线成像。在检测精度和计算时间方面,与知名变换和最近致力于此目的的工作进行了比较,其中所提出的变换在检测上述结构和准确定位其空间位置方面表现出色,即使在低质量图像中也是如此。
{"title":"Scale space Radon transform","authors":"D. Ziou, Nafaa Nacereddine, A. Goumeidane","doi":"10.1049/IPR2.12180","DOIUrl":"https://doi.org/10.1049/IPR2.12180","url":null,"abstract":"An extension of Radon transform by using a measure function capturing the user need is proposed. The new transform, called scale space Radon transform, is devoted to the case where the embedded shape in the image is not filiform. A case study is brought on a straight line and an ellipse where the SSRT behaviour in the scale space and in the presence of noise is deeply analyzed. In order to show the effectiveness of the proposed transform, the experiments have been carried out, first, on linear and elliptical structures generated synthetically subjected to strong altering conditions such blur and noise and then on structures images issued from real-world applications such as road traffic, satellite imagery and weld X-ray imaging. Comparisons in terms of detection accuracy and computational time with well-known transforms and recent work dedicated to this purpose are conducted, where the proposed transform shows an outstanding performance in detecting the above-mentioned structures and targeting accurately their spatial locations even in low-quality images.","PeriodicalId":13486,"journal":{"name":"IET Image Process.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73904437","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Remote sensing target tracking in satellite videos based on a variable-angle-adaptive Siamese network 基于变角度自适应Siamese网络的卫星视频遥感目标跟踪
Pub Date : 2021-03-17 DOI: 10.1049/IPR2.12170
Fukun Bi, Jiayi Sun, Jianhong Han, Yanping Wang, M. Bian
Funding information National Natural Science Foundation of China, Grant/Award Number: 61971006; Natural Science Foundation of Beijing Municipal, Grant/Award Number: 4192021 Abstract Remote sensing target tracking in satellite videos plays a key role in various fields. However, due to the complex backgrounds of satellite video sequences and many rotation changes of highly dynamic targets, typical target tracking methods for natural scenes cannot be used directly for such tasks, and their robustness and accuracy are difficult to guarantee. To address these problems, an algorithm is proposed for remote sensing target tracking in satellite videos based on a variable-angle-adaptive Siamese network (VAASN). Specifically, the method is based on the fully convolutional Siamese network (Siamese-FC). First, for the feature extraction stage, to reduce the impact of complex backgrounds, we present a new multifrequency feature representation method and introduce the octave convolution (OctConv) into the AlexNet architecture to adapt to the new feature representation. Then, for the tracking stage, to adapt to changes in target rotation, a variable-angle-adaptive module that uses a fast text detector with a single deep neural network (TextBoxes++) is introduced to extract angle information from the template frame and detection frames and performs angle consistency update operations on the detection frames. Finally, qualitative and quantitative experiments using satellite datasets show that the proposed method can improve tracking accuracy while achieving high efficiency.
国家自然科学基金资助/奖励号:61971006;摘要卫星视频中的遥感目标跟踪在多个领域发挥着关键作用。然而,由于卫星视频序列背景复杂,高动态目标旋转变化多,典型的自然场景目标跟踪方法不能直接用于此类任务,其鲁棒性和准确性难以保证。针对这些问题,提出了一种基于可变角度自适应Siamese网络(VAASN)的卫星视频遥感目标跟踪算法。具体来说,该方法是基于全卷积暹罗网络(Siamese- fc)。首先,在特征提取阶段,为了减少复杂背景的影响,我们提出了一种新的多频特征表示方法,并在AlexNet架构中引入了八度卷积(OctConv)来适应新的特征表示。然后,在跟踪阶段,为了适应目标旋转的变化,引入了一个可变角度自适应模块,该模块使用快速文本检测器和单个深度神经网络(textbox++),从模板帧和检测帧中提取角度信息,并对检测帧进行角度一致性更新操作。最后,利用卫星数据集进行了定性和定量实验,结果表明,该方法在提高跟踪精度的同时具有较高的效率。
{"title":"Remote sensing target tracking in satellite videos based on a variable-angle-adaptive Siamese network","authors":"Fukun Bi, Jiayi Sun, Jianhong Han, Yanping Wang, M. Bian","doi":"10.1049/IPR2.12170","DOIUrl":"https://doi.org/10.1049/IPR2.12170","url":null,"abstract":"Funding information National Natural Science Foundation of China, Grant/Award Number: 61971006; Natural Science Foundation of Beijing Municipal, Grant/Award Number: 4192021 Abstract Remote sensing target tracking in satellite videos plays a key role in various fields. However, due to the complex backgrounds of satellite video sequences and many rotation changes of highly dynamic targets, typical target tracking methods for natural scenes cannot be used directly for such tasks, and their robustness and accuracy are difficult to guarantee. To address these problems, an algorithm is proposed for remote sensing target tracking in satellite videos based on a variable-angle-adaptive Siamese network (VAASN). Specifically, the method is based on the fully convolutional Siamese network (Siamese-FC). First, for the feature extraction stage, to reduce the impact of complex backgrounds, we present a new multifrequency feature representation method and introduce the octave convolution (OctConv) into the AlexNet architecture to adapt to the new feature representation. Then, for the tracking stage, to adapt to changes in target rotation, a variable-angle-adaptive module that uses a fast text detector with a single deep neural network (TextBoxes++) is introduced to extract angle information from the template frame and detection frames and performs angle consistency update operations on the detection frames. Finally, qualitative and quantitative experiments using satellite datasets show that the proposed method can improve tracking accuracy while achieving high efficiency.","PeriodicalId":13486,"journal":{"name":"IET Image Process.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-03-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86497990","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
DuGAN: An effective framework for underwater image enhancement 一种有效的水下图像增强框架
Pub Date : 2021-03-14 DOI: 10.1049/IPR2.12172
Huiqing Zhang, Luyu Sun, Lifang Wu, Ke Gu
Underwater image enhancement is an important low-level vision task with much attention of community. Clear underwater images are helpful for underwater operations. However, raw underwater images often suffer from different types of distortions caused by the underwater environment. To solve these problems, this paper proposes an end-to-end dual generative adversarial network (DuGAN) for underwater image enhancement. The images processed by existing methods are taken as training samples for reference, and they are segmented into clear parts and unclear parts. Two discriminators are used to complete adversarial training toward different areas of images with different training strategies, respectively. The proposed method is able to output more pleasing images than reference images benefit by this framework. Meanwhile, to ensure the authenticity of the enhanced images, content loss, adversarial loss, and style loss are combined as loss function of our framework. This framework is easy to use, and the subjective and objective experiments show that excellent results are achieved compared to those methods mentioned in the literature.
水下图像增强是一项重要的低层次视觉任务,受到了广泛的关注。清晰的水下图像有助于水下作业。然而,原始的水下图像往往遭受不同类型的水下环境造成的畸变。为了解决这些问题,本文提出了一种端到端双生成对抗网络(DuGAN)用于水下图像增强。将现有方法处理后的图像作为训练样本进行参考,并将其分割为清晰部分和不清晰部分。使用两个判别器分别对图像的不同区域采用不同的训练策略完成对抗性训练。该方法能够输出比参考图像更令人满意的图像。同时,为了保证增强图像的真实性,我们将内容损失、对抗损失和风格损失作为我们框架的损失函数。该框架易于使用,主观和客观实验表明,与文献中提到的方法相比,该框架取得了优异的效果。
{"title":"DuGAN: An effective framework for underwater image enhancement","authors":"Huiqing Zhang, Luyu Sun, Lifang Wu, Ke Gu","doi":"10.1049/IPR2.12172","DOIUrl":"https://doi.org/10.1049/IPR2.12172","url":null,"abstract":"Underwater image enhancement is an important low-level vision task with much attention of community. Clear underwater images are helpful for underwater operations. However, raw underwater images often suffer from different types of distortions caused by the underwater environment. To solve these problems, this paper proposes an end-to-end dual generative adversarial network (DuGAN) for underwater image enhancement. The images processed by existing methods are taken as training samples for reference, and they are segmented into clear parts and unclear parts. Two discriminators are used to complete adversarial training toward different areas of images with different training strategies, respectively. The proposed method is able to output more pleasing images than reference images benefit by this framework. Meanwhile, to ensure the authenticity of the enhanced images, content loss, adversarial loss, and style loss are combined as loss function of our framework. This framework is easy to use, and the subjective and objective experiments show that excellent results are achieved compared to those methods mentioned in the literature.","PeriodicalId":13486,"journal":{"name":"IET Image Process.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80293871","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Towards accurate classification of skin cancer from dermatology images 从皮肤病学图像中准确分类皮肤癌
Pub Date : 2021-03-08 DOI: 10.1049/IPR2.12166
Anjali Gautam, B. Raman
Correspondence Anjali Gautam, Department of Information Technology, Indian Institute of Information Technology, Allahabad, Prayagraj, Uttar Pradesh, India. Email: anjaligautam@iiita.ac.in Abstract Skin cancer is the most well-known disease found in the individuals who are exposed to the Sun’s ultra-violet (UV) radiations. It is identified when skin tissues on the epidermis grow in an uncontrolled manner and appears to be of different colour than the normal skin tissues. This paper focuses on predicting the class of dermascopic images as benign and malignant. A new feature extraction method has been proposed to carry out this work which can extract relevant features from image texture. Local and gradient information from x and y directions of images has been utilized for feature extraction. After that images are classified using machine learning algorithms by using those extracted features. The efficacy of the proposed feature extraction method has been proved by conducting several experiments on the publicly available image dataset 2016 International Skin Imaging Collaboration (ISIC 2016). The classification results obtained by the method are also compared with state-of-the-art feature extraction methods which show that it performs better than others. The evaluation criteria used to obtain the results are accuracy, true positive rate (TPR) and false positive rate (FPR) where TPR and FPR are used for generating receiver operating characteristic curves.
Anjali Gautam,印度北方邦Prayagraj阿拉哈巴德印度信息技术学院信息技术系。摘要皮肤癌是在暴露于太阳紫外线(UV)辐射的个体中发现的最广为人知的疾病。当表皮上的皮肤组织以不受控制的方式生长,并且看起来与正常皮肤组织的颜色不同时,就可以识别出它。本文着重于预测皮肤镜图像的良性和恶性分类。为此,提出了一种新的特征提取方法,从图像纹理中提取相关特征。利用图像x和y方向的局部信息和梯度信息进行特征提取。然后使用机器学习算法利用这些提取的特征对图像进行分类。在2016年国际皮肤成像协作(ISIC 2016)公开的图像数据集上进行了多次实验,证明了所提出的特征提取方法的有效性。将该方法的分类结果与现有的特征提取方法进行了比较,结果表明该方法具有较好的分类效果。获得结果的评价标准为准确率、真阳性率(TPR)和假阳性率(FPR),其中用真阳性率和假阳性率生成受试者工作特性曲线。
{"title":"Towards accurate classification of skin cancer from dermatology images","authors":"Anjali Gautam, B. Raman","doi":"10.1049/IPR2.12166","DOIUrl":"https://doi.org/10.1049/IPR2.12166","url":null,"abstract":"Correspondence Anjali Gautam, Department of Information Technology, Indian Institute of Information Technology, Allahabad, Prayagraj, Uttar Pradesh, India. Email: anjaligautam@iiita.ac.in Abstract Skin cancer is the most well-known disease found in the individuals who are exposed to the Sun’s ultra-violet (UV) radiations. It is identified when skin tissues on the epidermis grow in an uncontrolled manner and appears to be of different colour than the normal skin tissues. This paper focuses on predicting the class of dermascopic images as benign and malignant. A new feature extraction method has been proposed to carry out this work which can extract relevant features from image texture. Local and gradient information from x and y directions of images has been utilized for feature extraction. After that images are classified using machine learning algorithms by using those extracted features. The efficacy of the proposed feature extraction method has been proved by conducting several experiments on the publicly available image dataset 2016 International Skin Imaging Collaboration (ISIC 2016). The classification results obtained by the method are also compared with state-of-the-art feature extraction methods which show that it performs better than others. The evaluation criteria used to obtain the results are accuracy, true positive rate (TPR) and false positive rate (FPR) where TPR and FPR are used for generating receiver operating characteristic curves.","PeriodicalId":13486,"journal":{"name":"IET Image Process.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-03-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77876219","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Lemon-YOLO: An efficient object detection method for lemons in the natural environment 柠檬- yolo:一种针对自然环境中柠檬的高效目标检测方法
Pub Date : 2021-03-08 DOI: 10.1049/IPR2.12171
Guojin Li, Xiaojie Huang, Jiaoyan Ai, Zeren Yi, Wei Xie
Efficient Intelligent detection is a key technology in automatic harvesting robots. How-ever, citrus detection is still a challenging task because of varying illumination, random occlusion and colour similarity between fruits and leaves in natural conditions. In this paper, a detection method called Lemon-YOLO (L-YOLO) is proposed to improve the accuracy and real-time performance of lemon detection in the natural environment. The SE_ResGNet34 network is designed to replace DarkNet53 network in YOLOv3 algorithm as a new backbone of feature extraction. It can enhance the propagation of features, and needs less parameter, which helps to achieve higher accuracy and speed. Moreover, the SE_ResNet module is added to the detection block, to improve the quality of representa-tions produced from the network by strengthening the convolutional features of channels. The experimental results show that the proposed L-YOLO has an average accuracy(AP) of 96.28% and a detection speed of 106 frames per second (FPS) on the lemon test set, which is 5.68% and 28 FPS higher than the YOLOv3, respectively. The results indicate that the L-YOLO method has superior detection performance. It can recognize and locate lemons in the natural environment more efficiently, providing technical support for the machine’s picking lemon and other fruits.
高效智能检测是自动采收机器人的关键技术。然而,柑橘的检测仍然是一项具有挑战性的任务,因为在自然条件下,水果和叶子之间存在不同的光照、随机遮挡和颜色相似性。为了提高自然环境中柠檬检测的准确性和实时性,本文提出了一种柠檬- yolo (L-YOLO)检测方法。SE_ResGNet34网络旨在取代YOLOv3算法中的DarkNet53网络,成为新的特征提取骨干。它可以增强特征的传播,并且需要较少的参数,有助于达到更高的精度和速度。此外,在检测块中增加了SE_ResNet模块,通过增强通道的卷积特征来提高网络产生的表示的质量。实验结果表明,本文提出的L-YOLO在柠檬测试集上的平均准确率(AP)为96.28%,检测速度为106帧/秒(FPS),分别比YOLOv3高5.68%和28 FPS。结果表明,L-YOLO方法具有较好的检测性能。它可以更高效地识别和定位自然环境中的柠檬,为机器采摘柠檬等水果提供技术支持。
{"title":"Lemon-YOLO: An efficient object detection method for lemons in the natural environment","authors":"Guojin Li, Xiaojie Huang, Jiaoyan Ai, Zeren Yi, Wei Xie","doi":"10.1049/IPR2.12171","DOIUrl":"https://doi.org/10.1049/IPR2.12171","url":null,"abstract":"Efficient Intelligent detection is a key technology in automatic harvesting robots. How-ever, citrus detection is still a challenging task because of varying illumination, random occlusion and colour similarity between fruits and leaves in natural conditions. In this paper, a detection method called Lemon-YOLO (L-YOLO) is proposed to improve the accuracy and real-time performance of lemon detection in the natural environment. The SE_ResGNet34 network is designed to replace DarkNet53 network in YOLOv3 algorithm as a new backbone of feature extraction. It can enhance the propagation of features, and needs less parameter, which helps to achieve higher accuracy and speed. Moreover, the SE_ResNet module is added to the detection block, to improve the quality of representa-tions produced from the network by strengthening the convolutional features of channels. The experimental results show that the proposed L-YOLO has an average accuracy(AP) of 96.28% and a detection speed of 106 frames per second (FPS) on the lemon test set, which is 5.68% and 28 FPS higher than the YOLOv3, respectively. The results indicate that the L-YOLO method has superior detection performance. It can recognize and locate lemons in the natural environment more efficiently, providing technical support for the machine’s picking lemon and other fruits.","PeriodicalId":13486,"journal":{"name":"IET Image Process.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-03-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86854008","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Weakly supervised salient object detection via double object proposals guidance 弱监督显著目标检测采用双目标建议指导
Pub Date : 2021-03-04 DOI: 10.1049/IPR2.12164
Zhiheng Zhou, Yongfan Guo, Ming Dai, Junchu Huang, Xiangwei Li
Funding information National Natural Science Foundation of China, Grant/Award Number: 61871188; National Key R&D Program of China, Grant/Award Number: 2018YFC0309400; Guangzhou city science and technology research projects, Grant/Award Number: 201902020008 Abstract The weakly supervised methods for salient object detection are attractive, since they greatly release the burden of annotating time-consuming pixel-wise masks. However, the imagelevel annotations utilized by current weakly supervised salient object detection models are too weak to provide sufficient supervision for this dense prediction task. To this end, a weakly supervised salient object detection method is proposed via double object proposals guidance, which is generated under the supervision of double bounding boxes annotations. With the double object proposals, the authors’ method is capable of capturing both accurate but incomplete salient foreground and background information, which contributes to generating saliency maps with uniformly highlighted saliency regions and effectively suppressed background. In addition, an unsupervised salient object segmentation method is proposed, taking advantage of the non-parametric statistical active contour model (NSACM), for segmenting salient objects with complete and compact boundaries. Experiments on five benchmark datasets show that the authors’ weakly supervised salient object detection approach consistently outperforms other weakly supervised and unsupervised methods by a considerable margin, and even has comparable performance to the fully supervised ones.
国家自然科学基金资助/奖励号:61871188;国家重点科技发展计划资助/奖励号:2018YFC0309400;摘要基于弱监督的显著目标检测方法具有很大的吸引力,因为它极大地减轻了标注耗时的像素掩模的负担。然而,现有的弱监督显著目标检测模型所使用的图像级标注太弱,无法为这种密集的预测任务提供足够的监督。为此,提出了一种弱监督显著目标检测方法,该方法是在双边界框标注的监督下生成的双目标建议指导。采用双目标方法,作者的方法能够捕获准确但不完整的显著前景和背景信息,有助于生成均匀突出突出区域和有效抑制背景的显著性地图。此外,利用非参数统计活动轮廓模型(NSACM),提出了一种无监督显著目标分割方法,用于分割边界完备紧凑的显著目标。在五个基准数据集上的实验表明,作者的弱监督显著目标检测方法始终优于其他弱监督和无监督方法,甚至与完全监督方法的性能相当。
{"title":"Weakly supervised salient object detection via double object proposals guidance","authors":"Zhiheng Zhou, Yongfan Guo, Ming Dai, Junchu Huang, Xiangwei Li","doi":"10.1049/IPR2.12164","DOIUrl":"https://doi.org/10.1049/IPR2.12164","url":null,"abstract":"Funding information National Natural Science Foundation of China, Grant/Award Number: 61871188; National Key R&D Program of China, Grant/Award Number: 2018YFC0309400; Guangzhou city science and technology research projects, Grant/Award Number: 201902020008 Abstract The weakly supervised methods for salient object detection are attractive, since they greatly release the burden of annotating time-consuming pixel-wise masks. However, the imagelevel annotations utilized by current weakly supervised salient object detection models are too weak to provide sufficient supervision for this dense prediction task. To this end, a weakly supervised salient object detection method is proposed via double object proposals guidance, which is generated under the supervision of double bounding boxes annotations. With the double object proposals, the authors’ method is capable of capturing both accurate but incomplete salient foreground and background information, which contributes to generating saliency maps with uniformly highlighted saliency regions and effectively suppressed background. In addition, an unsupervised salient object segmentation method is proposed, taking advantage of the non-parametric statistical active contour model (NSACM), for segmenting salient objects with complete and compact boundaries. Experiments on five benchmark datasets show that the authors’ weakly supervised salient object detection approach consistently outperforms other weakly supervised and unsupervised methods by a considerable margin, and even has comparable performance to the fully supervised ones.","PeriodicalId":13486,"journal":{"name":"IET Image Process.","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2021-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90891966","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
IET Image Process.
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1