首页 > 最新文献

International Journal of Image and Graphics最新文献

英文 中文
Research on robust digital watermarking based on reversible information hiding 基于可逆信息隐藏的鲁棒数字水印研究
IF 1.6 Q4 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2023-08-17 DOI: 10.1142/s0219467825500354
Zhijing Gao, Weilin Qiu, Ren Wenqi, Xiao Yan
{"title":"Research on robust digital watermarking based on reversible information hiding","authors":"Zhijing Gao, Weilin Qiu, Ren Wenqi, Xiao Yan","doi":"10.1142/s0219467825500354","DOIUrl":"https://doi.org/10.1142/s0219467825500354","url":null,"abstract":"","PeriodicalId":44688,"journal":{"name":"International Journal of Image and Graphics","volume":" ","pages":""},"PeriodicalIF":1.6,"publicationDate":"2023-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49136822","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimal Classification Model for Text Detection and Recognition in Video Frames 视频帧文本检测与识别的最优分类模型
IF 1.6 Q4 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2023-08-04 DOI: 10.1142/s0219467825500147
Laxmikant Eshwarappa, G. G. Rajput
Currently, the identification of text from video frames and normal scene images has got amplified awareness amongst analysts owing to its diverse challenges and complexities. Owing to a lower resolution, composite backdrop, blurring effect, color, diverse fonts, alternate textual placement among panels of photos and videos, etc., text identification is becoming complicated. This paper suggests a novel method for identifying texts from video with five stages. Initially, “video-to-frame conversion”, is done during pre-processing. Further, text region verification is performed and keyframes are recognized using CNN. Then, improved candidate text block extraction is carried out using MSER. Subsequently, “DCT features, improved distance map features, and constant gradient-based features” are extracted. These characteristics subsequently provided “Long Short-Term Memory (LSTM)” for detection. Finally, OCR is done to recognize the texts in the image. Particularly, the Self-Improved Bald Eagle Search (SI-BESO) algorithm is used to adjust the LSTM weights. Finally, the superiority of the SI-BESO-based technique over many other techniques is demonstrated.
目前,从视频帧和正常场景图像中识别文本由于其多样的挑战和复杂性而在分析师中得到了广泛的关注。由于较低的分辨率、复合背景、模糊效果、颜色、不同的字体、照片和视频面板之间的交替文本位置等,文本识别变得越来越复杂。本文提出了一种从视频中识别文本的新方法,该方法分为五个阶段。最初,“视频到帧的转换”是在预处理过程中完成的。此外,使用CNN执行文本区域验证并识别关键帧。然后,使用MSER进行改进的候选文本块提取。随后,提取了“DCT特征、改进的距离图特征和基于恒定梯度的特征”。这些特征随后为检测提供了“长短期记忆(LSTM)”。最后,对图像中的文本进行OCR识别。特别地,使用自改进的秃鹰搜索(SI-BESO)算法来调整LSTM权重。最后,证明了基于SI BESO的技术相对于许多其他技术的优越性。
{"title":"Optimal Classification Model for Text Detection and Recognition in Video Frames","authors":"Laxmikant Eshwarappa, G. G. Rajput","doi":"10.1142/s0219467825500147","DOIUrl":"https://doi.org/10.1142/s0219467825500147","url":null,"abstract":"Currently, the identification of text from video frames and normal scene images has got amplified awareness amongst analysts owing to its diverse challenges and complexities. Owing to a lower resolution, composite backdrop, blurring effect, color, diverse fonts, alternate textual placement among panels of photos and videos, etc., text identification is becoming complicated. This paper suggests a novel method for identifying texts from video with five stages. Initially, “video-to-frame conversion”, is done during pre-processing. Further, text region verification is performed and keyframes are recognized using CNN. Then, improved candidate text block extraction is carried out using MSER. Subsequently, “DCT features, improved distance map features, and constant gradient-based features” are extracted. These characteristics subsequently provided “Long Short-Term Memory (LSTM)” for detection. Finally, OCR is done to recognize the texts in the image. Particularly, the Self-Improved Bald Eagle Search (SI-BESO) algorithm is used to adjust the LSTM weights. Finally, the superiority of the SI-BESO-based technique over many other techniques is demonstrated.","PeriodicalId":44688,"journal":{"name":"International Journal of Image and Graphics","volume":" ","pages":""},"PeriodicalIF":1.6,"publicationDate":"2023-08-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48648435","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Jeap-BiLSTM Neural Network for Action Recognition 用于动作识别的Jeap-BiLSTM神经网络
IF 1.6 Q4 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2023-08-03 DOI: 10.1142/s0219467825500184
Lunzheng Tan, Yanfei Liu, Li-min Xia, Shangsheng Chen, Zhanben Zhou
Human action recognition in videos is an important task in computer vision with applications in fields such as surveillance, human–computer interaction, and sports analysis. However, it is a challenging task due to the complex background changes and redundancy of long-term video information. In this paper, we propose a novel bi-directional long short-term memory method with attention pooling based on joint motion and difference entropy (JEAP-BiLSTM) to address these challenges. To obtain discriminative features, we introduce a joint entropy map that measures both the entropy of motion and the entropy of change. The Bi-LSTM method is then applied to capture visual and temporal associations in both forward and backward directions, enabling efficient capture of long-term temporal correlation. Furthermore, attention pooling is used to highlight the region of interest and to mitigate the effects of background changes in video information. Experiments on the UCF101 and HMDB51 datasets demonstrate that the proposed JEAP-BiLSTM method achieves recognition rates of 96.4% and 75.2%, respectively, outperforming existing methods. Our proposed method makes significant contributions to the field of human action recognition by effectively capturing both spatial and temporal patterns in videos, addressing background changes, and achieving state-of-the-art performance.
视频中的人类动作识别是计算机视觉中的一项重要任务,在监控、人机交互和体育分析等领域都有应用。然而,由于复杂的背景变化和长期视频信息的冗余,这是一项具有挑战性的任务。在本文中,我们提出了一种新的基于联合运动和差分熵的注意力池双向长短期记忆方法(JEAP BiLSTM)来应对这些挑战。为了获得判别特征,我们引入了一个测量运动熵和变化熵的联合熵图。然后,Bi-LSTM方法被应用于捕捉前向和后向的视觉和时间关联,从而能够有效地捕捉长期时间相关性。此外,注意力集中用于突出感兴趣的区域并减轻视频信息中背景变化的影响。在UCF101和HMDB51数据集上的实验表明,所提出的JEAP BiLSTM方法的识别率分别为96.4%和75.2%,优于现有方法。我们提出的方法通过有效捕捉视频中的空间和时间模式,解决背景变化,并实现最先进的性能,为人类动作识别领域做出了重大贡献。
{"title":"A Jeap-BiLSTM Neural Network for Action Recognition","authors":"Lunzheng Tan, Yanfei Liu, Li-min Xia, Shangsheng Chen, Zhanben Zhou","doi":"10.1142/s0219467825500184","DOIUrl":"https://doi.org/10.1142/s0219467825500184","url":null,"abstract":"Human action recognition in videos is an important task in computer vision with applications in fields such as surveillance, human–computer interaction, and sports analysis. However, it is a challenging task due to the complex background changes and redundancy of long-term video information. In this paper, we propose a novel bi-directional long short-term memory method with attention pooling based on joint motion and difference entropy (JEAP-BiLSTM) to address these challenges. To obtain discriminative features, we introduce a joint entropy map that measures both the entropy of motion and the entropy of change. The Bi-LSTM method is then applied to capture visual and temporal associations in both forward and backward directions, enabling efficient capture of long-term temporal correlation. Furthermore, attention pooling is used to highlight the region of interest and to mitigate the effects of background changes in video information. Experiments on the UCF101 and HMDB51 datasets demonstrate that the proposed JEAP-BiLSTM method achieves recognition rates of 96.4% and 75.2%, respectively, outperforming existing methods. Our proposed method makes significant contributions to the field of human action recognition by effectively capturing both spatial and temporal patterns in videos, addressing background changes, and achieving state-of-the-art performance.","PeriodicalId":44688,"journal":{"name":"International Journal of Image and Graphics","volume":" ","pages":""},"PeriodicalIF":1.6,"publicationDate":"2023-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46374210","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Survey on Epileptic Seizure Detection on Varied Machine Learning Algorithms 基于不同机器学习算法的癫痫发作检测研究综述
IF 1.6 Q4 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2023-08-03 DOI: 10.1142/s0219467825500135
Nusrat Fatma, Pawan Singh, M. K. Siddiqui
Epilepsy is an unavoidable major persistent and critical neurological disorder that influences the human brain. Moreover, this is apparently distinguished via its recurrent malicious seizures. A seizure is a phase of synchronous, abnormal innervations of a neuron’s population which might last from seconds to a few minutes. In addition, epileptic seizures are transient occurrences of complete or partial irregular unintentional body movements that combine with consciousness loss. As epileptic seizures rarely occurred in each patient, their effects based on physical communications, social interactions, and patients’ emotions are considered, and treatment and diagnosis are undergone with crucial implications. Therefore, this survey reviews 65 research papers and states an important analysis on various machine-learning approaches adopted in each paper. The analysis of different features considered in each work is also done. This survey offers a comprehensive study on performance attainment in each contribution. Furthermore, the maximum performance attained by the works and the datasets used in each work is also examined. The analysis on features and the simulation tools used in each contribution is examined. At the end, the survey expanded with different research gaps and their problem which is beneficial to the researchers for promoting advanced future works on epileptic seizure detection.
癫痫是一种不可避免的影响人类大脑的重大、持续和严重的神经系统疾病。此外,这显然是通过其反复发作的恶意发作来区分的。癫痫发作是一个同步的阶段,神经元群的异常神经支配可能持续几秒到几分钟。此外,癫痫发作是完全或部分不规则的无意识身体运动的短暂发生,并伴有意识丧失。由于癫痫发作很少发生在每个患者身上,因此基于身体交流、社会互动和患者情绪的影响被考虑在内,并且进行了具有重要意义的治疗和诊断。因此,本调查回顾了65篇研究论文,并对每篇论文中采用的各种机器学习方法进行了重要分析。并对各作品所考虑的不同特点进行了分析。这项调查提供了对每个贡献的绩效成就的全面研究。此外,还检查了作品和每个作品中使用的数据集所获得的最大性能。分析了每个贡献的特征和使用的仿真工具。最后,对不同的研究空白和存在的问题进行了扩展,有利于研究人员推进未来癫痫发作检测工作。
{"title":"Survey on Epileptic Seizure Detection on Varied Machine Learning Algorithms","authors":"Nusrat Fatma, Pawan Singh, M. K. Siddiqui","doi":"10.1142/s0219467825500135","DOIUrl":"https://doi.org/10.1142/s0219467825500135","url":null,"abstract":"Epilepsy is an unavoidable major persistent and critical neurological disorder that influences the human brain. Moreover, this is apparently distinguished via its recurrent malicious seizures. A seizure is a phase of synchronous, abnormal innervations of a neuron’s population which might last from seconds to a few minutes. In addition, epileptic seizures are transient occurrences of complete or partial irregular unintentional body movements that combine with consciousness loss. As epileptic seizures rarely occurred in each patient, their effects based on physical communications, social interactions, and patients’ emotions are considered, and treatment and diagnosis are undergone with crucial implications. Therefore, this survey reviews 65 research papers and states an important analysis on various machine-learning approaches adopted in each paper. The analysis of different features considered in each work is also done. This survey offers a comprehensive study on performance attainment in each contribution. Furthermore, the maximum performance attained by the works and the datasets used in each work is also examined. The analysis on features and the simulation tools used in each contribution is examined. At the end, the survey expanded with different research gaps and their problem which is beneficial to the researchers for promoting advanced future works on epileptic seizure detection.","PeriodicalId":44688,"journal":{"name":"International Journal of Image and Graphics","volume":" ","pages":""},"PeriodicalIF":1.6,"publicationDate":"2023-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44591500","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Research on Printmaking Image Classification and Creation Based on Convolutional Neural Network 基于卷积神经网络的版画图像分类与创作研究
IF 1.6 Q4 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2023-08-03 DOI: 10.1142/s0219467825500196
Kai Pan, Hongyan Chi
As an important form of expression in modern civilization art, printmaking has a rich variety of types and a prominent sense of artistic hierarchy. Therefore, printmaking is highly favored around the world due to its unique artistic characteristics. Classifying print types through image feature elements will improve people’s understanding of print creation. Convolutional neural networks (CNNs) have good application effects in the field of image classification, so CNN is used for printmaking analysis. Considering that the classification effect of the traditional convolutional neural image classification model is easily affected by the activation function, the T-ReLU activation function is introduced. By utilizing adjustable parameters to enhance the soft saturation characteristics of the model and avoid gradient vanishing, a T-ReLU convolutional model is constructed. A better convolutional image classification model is proposed based on the T-ReLU convolutional model, taking into account the issue of subpar multi-level feature fusion in deep convolutional image classification models. Utilize normalization to analyze visual input, an eleven-layer convolutional network with residual units in the convolutional layer, and cascading thinking to fuse convolutional network defects. The performance test results showed that in the data test of different styles of artificial prints, the GT-ReLU model can obtain the best image classification accuracy, and the image classification accuracy rate is 0.978. The GT-ReLU model maintains a classification accuracy above 94.4% in the multi-dataset test classification performance test, which is higher than that of other image classification models. For the use of visual processing technology in the field of classifying prints, the research content provides good reference value.
版画作为现代文明艺术的重要表现形式,种类丰富,艺术层次感突出。因此,版画以其独特的艺术特征在世界范围内备受青睐。通过图像特征元素对版画类型进行分类,可以提高人们对版画创作的理解。卷积神经网络(CNN)在图像分类领域有很好的应用效果,因此将CNN用于版画分析。考虑到传统卷积神经图像分类模型的分类效果容易受到激活函数的影响,引入T-ReLU激活函数。利用可调参数增强模型的软饱和特性,避免梯度消失,构造了T-ReLU卷积模型。针对深度卷积图像分类模型中多级特征融合不足的问题,在T-ReLU卷积模型的基础上,提出了一种更好的卷积图像分类模型。利用归一化分析视觉输入,利用卷积层残差单元的11层卷积网络,利用级联思维融合卷积网络缺陷。性能测试结果表明,在不同款式的人造指纹数据测试中,GT-ReLU模型能获得最佳的图像分类精度,图像分类准确率为0.978。在多数据集测试分类性能测试中,GT-ReLU模型的分类准确率保持在94.4%以上,高于其他图像分类模型。对于视觉处理技术在印刷品分类领域的应用,研究内容具有很好的参考价值。
{"title":"Research on Printmaking Image Classification and Creation Based on Convolutional Neural Network","authors":"Kai Pan, Hongyan Chi","doi":"10.1142/s0219467825500196","DOIUrl":"https://doi.org/10.1142/s0219467825500196","url":null,"abstract":"As an important form of expression in modern civilization art, printmaking has a rich variety of types and a prominent sense of artistic hierarchy. Therefore, printmaking is highly favored around the world due to its unique artistic characteristics. Classifying print types through image feature elements will improve people’s understanding of print creation. Convolutional neural networks (CNNs) have good application effects in the field of image classification, so CNN is used for printmaking analysis. Considering that the classification effect of the traditional convolutional neural image classification model is easily affected by the activation function, the T-ReLU activation function is introduced. By utilizing adjustable parameters to enhance the soft saturation characteristics of the model and avoid gradient vanishing, a T-ReLU convolutional model is constructed. A better convolutional image classification model is proposed based on the T-ReLU convolutional model, taking into account the issue of subpar multi-level feature fusion in deep convolutional image classification models. Utilize normalization to analyze visual input, an eleven-layer convolutional network with residual units in the convolutional layer, and cascading thinking to fuse convolutional network defects. The performance test results showed that in the data test of different styles of artificial prints, the GT-ReLU model can obtain the best image classification accuracy, and the image classification accuracy rate is 0.978. The GT-ReLU model maintains a classification accuracy above 94.4% in the multi-dataset test classification performance test, which is higher than that of other image classification models. For the use of visual processing technology in the field of classifying prints, the research content provides good reference value.","PeriodicalId":44688,"journal":{"name":"International Journal of Image and Graphics","volume":" ","pages":""},"PeriodicalIF":1.6,"publicationDate":"2023-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48183056","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimization with Deep Learning Classifier-Based Foliar Disease Classification in Apple Trees Using IoT Network 基于深度学习分类器的物联网苹果树叶病分类优化
IF 1.6 Q4 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2023-08-03 DOI: 10.1142/s0219467825500159
K. Sameera, P. Swarnalatha
The development of any country is influenced by the growth in the agriculture sector. The prevalence of pests and diseases in plants affects the productivity of any agricultural product. Early diagnosis of the disease can substantially decrease the effort and the fund required for disease management. The Internet of Things (IoT) provides a framework for offering solutions for automatic farming. This paper devises an automated detection technique for foliar disease classification in apple trees using an IoT network. Here, classification is performed using a hybrid classifier, which utilizes the Deep Residual Network (DRN) and Deep [Formula: see text] Network (DQN). A new Adaptive Tunicate Swarm Sine–Cosine Algorithm (TSSCA) is used for modifying the learning parameters as well as the weights of the proposed hybrid classifier. The TSSCA is developed by adaptively changing the navigation foraging behavior of the tunicates obtained from the Tunicate Swarm Algorithm (TSA) in accordance with the Sine–Cosine Algorithm  (SCA). The outputs obtained from the Adaptive TSSCA-based DRN and Adaptive TSSCA-based DQN are merged using cosine similarity measure for detecting the foliar disease. The Plant Pathology 2020 — FGVC7 dataset is utilized for the experimental process to determine accuracy, sensitivity, specificity and energy and we achieved the values of 98.36%, 98.58%, 96.32% and 0.413 J, respectively.
任何国家的发展都受到农业部门增长的影响。植物病虫害的流行影响着任何农产品的生产力。疾病的早期诊断可以大大减少疾病管理所需的努力和资金。物联网(IoT)为提供自动化农业解决方案提供了一个框架。本文设计了一种基于物联网网络的苹果树叶面病害分类自动检测技术。在这里,使用混合分类器进行分类,该分类器利用了深度残差网络(DRN)和深度[公式:见文本]网络(DQN)。采用一种新的自适应束状虫群正弦余弦算法(TSSCA)来修改混合分类器的学习参数和权重。TSSCA是根据正弦余弦算法(SCA)自适应改变由被囊动物群算法(TSA)得到的被囊动物的导航觅食行为而发展起来的。将基于自适应tssca的DRN和基于自适应tssca的DQN的输出用余弦相似度度量合并,用于叶面病害检测。实验过程使用Plant Pathology 2020 - FGVC7数据集来确定准确性、灵敏度、特异性和能量,我们分别获得了98.36%、98.58%、96.32%和0.413 J的值。
{"title":"Optimization with Deep Learning Classifier-Based Foliar Disease Classification in Apple Trees Using IoT Network","authors":"K. Sameera, P. Swarnalatha","doi":"10.1142/s0219467825500159","DOIUrl":"https://doi.org/10.1142/s0219467825500159","url":null,"abstract":"The development of any country is influenced by the growth in the agriculture sector. The prevalence of pests and diseases in plants affects the productivity of any agricultural product. Early diagnosis of the disease can substantially decrease the effort and the fund required for disease management. The Internet of Things (IoT) provides a framework for offering solutions for automatic farming. This paper devises an automated detection technique for foliar disease classification in apple trees using an IoT network. Here, classification is performed using a hybrid classifier, which utilizes the Deep Residual Network (DRN) and Deep [Formula: see text] Network (DQN). A new Adaptive Tunicate Swarm Sine–Cosine Algorithm (TSSCA) is used for modifying the learning parameters as well as the weights of the proposed hybrid classifier. The TSSCA is developed by adaptively changing the navigation foraging behavior of the tunicates obtained from the Tunicate Swarm Algorithm (TSA) in accordance with the Sine–Cosine Algorithm  (SCA). The outputs obtained from the Adaptive TSSCA-based DRN and Adaptive TSSCA-based DQN are merged using cosine similarity measure for detecting the foliar disease. The Plant Pathology 2020 — FGVC7 dataset is utilized for the experimental process to determine accuracy, sensitivity, specificity and energy and we achieved the values of 98.36%, 98.58%, 96.32% and 0.413 J, respectively.","PeriodicalId":44688,"journal":{"name":"International Journal of Image and Graphics","volume":" ","pages":""},"PeriodicalIF":1.6,"publicationDate":"2023-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45955735","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Product Image Recommendation with Transformer Model Using Deep Reinforcement Learning 基于深度强化学习的变压器模型产品图像推荐
IF 1.6 Q4 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2023-08-03 DOI: 10.1142/s0219467825500202
Yuan Liu
A product image recommendation algorithm with transformer model using deep reinforcement learning is proposed. First, the product image recommendation architecture is designed to collect users’ historical product image clicking behaviors through the log information layer. The recommendation strategy layer uses collaborative filtering algorithm to calculate users’ long-term shopping interest and gated recurrent unit to calculate users’ short-term shopping interest, and predicts users’ long-term and short-term interest output based on users’ positive and negative feedback sequences. Second, the prediction results are fed into the transformer model for content planning to make the data format more suitable for subsequent content recommendation. Finally, the planning results of the transformer model are input to Deep Q-Leaning Network to obtain product image recommendation sequences under the learning of this network, and the results are transmitted to the data result layer, and finally presented to users through the presentation layer. The results show that the recommendation results of the proposed algorithm are consistent with the user’s browsing records. The average accuracy of product image recommendation is 97.1%, the maximum recommended time is 1.0[Formula: see text]s, the coverage and satisfaction are high, and the practical application effect is good. It can recommend more suitable products for users and promote the further development of e-commerce.
提出了一种基于深度强化学习的变压器模型产品图像推荐算法。首先,设计了产品图片推荐架构,通过日志信息层收集用户的历史产品图片点击行为。推荐策略层使用协同过滤算法计算用户的长期购物兴趣,使用门控递归单元计算用户的短期购物兴趣,并根据用户的正负反馈序列预测用户的长期和短期兴趣输出。其次,将预测结果馈送到用于内容规划的转换器模型中,以使数据格式更适合于后续的内容推荐。最后,将transformer模型的规划结果输入到深度Q学习网络,在该网络的学习下获得产品图像推荐序列,并将结果传输到数据结果层,最终通过表示层呈现给用户。结果表明,该算法的推荐结果与用户的浏览记录一致。产品图片推荐平均准确率为97.1%,最长推荐时间为1.0[公式:见正文]s,覆盖率和满意度较高,实际应用效果良好。它可以为用户推荐更合适的产品,促进电子商务的进一步发展。
{"title":"Product Image Recommendation with Transformer Model Using Deep Reinforcement Learning","authors":"Yuan Liu","doi":"10.1142/s0219467825500202","DOIUrl":"https://doi.org/10.1142/s0219467825500202","url":null,"abstract":"A product image recommendation algorithm with transformer model using deep reinforcement learning is proposed. First, the product image recommendation architecture is designed to collect users’ historical product image clicking behaviors through the log information layer. The recommendation strategy layer uses collaborative filtering algorithm to calculate users’ long-term shopping interest and gated recurrent unit to calculate users’ short-term shopping interest, and predicts users’ long-term and short-term interest output based on users’ positive and negative feedback sequences. Second, the prediction results are fed into the transformer model for content planning to make the data format more suitable for subsequent content recommendation. Finally, the planning results of the transformer model are input to Deep Q-Leaning Network to obtain product image recommendation sequences under the learning of this network, and the results are transmitted to the data result layer, and finally presented to users through the presentation layer. The results show that the recommendation results of the proposed algorithm are consistent with the user’s browsing records. The average accuracy of product image recommendation is 97.1%, the maximum recommended time is 1.0[Formula: see text]s, the coverage and satisfaction are high, and the practical application effect is good. It can recommend more suitable products for users and promote the further development of e-commerce.","PeriodicalId":44688,"journal":{"name":"International Journal of Image and Graphics","volume":" ","pages":""},"PeriodicalIF":1.6,"publicationDate":"2023-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43365571","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improved Invasive Weed Social Ski-Driver Optimization-Based Deep Convolution Neural Network for Diabetic Retinopathy Classification 改进的基于深度卷积神经网络的有创杂草社交滑雪驱动程序优化用于糖尿病视网膜病变分类
IF 1.6 Q4 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2023-08-03 DOI: 10.1142/s0219467825500123
Padmanayana Bhat, B. Anoop
The eye-related problem of diabetes is called diabetic retinopathy (DR), which is the main factor contributing to visual loss. This research develops an enhanced deep model for DR classification. Here, deep convolutional neural network (Deep CNN) is trained with the improved invasive weed social ski-driver optimization (IISSDO), which is generated by fusing improved invasive weed optimization (IIWO) and social ski-driver (SSD). The IISSDO-based Deep CNN classifies DR severity into normal, mild, non-proliferative DR (NPDR), moderate NPDR, severe NPDR, and proliferative. Initially, a type 2 fuzzy and cuckoo search (T2FCS) filter performs pre-processing and the quality of the data is improved by data augmentation. The lesion is then divided using DeepJoint segmentation. Then, the Deep CNN determines the DR. The analysis uses the Indian DR image database. The IISSDO-based Deep CNN has the highest accuracy, sensitivity, and specificity of 96.566%, 96.773%, and 96.517%, respectively.
糖尿病的眼睛相关问题被称为糖尿病视网膜病变(DR),这是导致视力丧失的主要因素。本研究开发了一种用于DR分类的增强深度模型。在这里,深度卷积神经网络(deep CNN)使用改进的入侵杂草社交滑雪驱动程序优化(IISSDO)进行训练,该优化是通过融合改进的入侵除草优化(IIWO)和社交滑雪驱动(SSD)生成的。基于IISSDO的Deep CNN将DR的严重程度分为正常、轻度、非增殖性DR(NPDR)、中度NPDR、重度NPDR和增殖性。最初,2型模糊杜鹃搜索(T2FCS)滤波器进行预处理,并通过数据扩充来提高数据的质量。然后使用DeepJoint分割对病变进行分割。然后,深度CNN确定DR。分析使用印度DR图像数据库。基于IISSDO的深度CNN具有最高的准确性、敏感性和特异性,分别为96.566%、96.773%和96.517%。
{"title":"Improved Invasive Weed Social Ski-Driver Optimization-Based Deep Convolution Neural Network for Diabetic Retinopathy Classification","authors":"Padmanayana Bhat, B. Anoop","doi":"10.1142/s0219467825500123","DOIUrl":"https://doi.org/10.1142/s0219467825500123","url":null,"abstract":"The eye-related problem of diabetes is called diabetic retinopathy (DR), which is the main factor contributing to visual loss. This research develops an enhanced deep model for DR classification. Here, deep convolutional neural network (Deep CNN) is trained with the improved invasive weed social ski-driver optimization (IISSDO), which is generated by fusing improved invasive weed optimization (IIWO) and social ski-driver (SSD). The IISSDO-based Deep CNN classifies DR severity into normal, mild, non-proliferative DR (NPDR), moderate NPDR, severe NPDR, and proliferative. Initially, a type 2 fuzzy and cuckoo search (T2FCS) filter performs pre-processing and the quality of the data is improved by data augmentation. The lesion is then divided using DeepJoint segmentation. Then, the Deep CNN determines the DR. The analysis uses the Indian DR image database. The IISSDO-based Deep CNN has the highest accuracy, sensitivity, and specificity of 96.566%, 96.773%, and 96.517%, respectively.","PeriodicalId":44688,"journal":{"name":"International Journal of Image and Graphics","volume":" ","pages":""},"PeriodicalIF":1.6,"publicationDate":"2023-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41622319","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimal Multisecret Image Sharing Using Lightweight Visual Sign-Cryptography Scheme With Optimal Key Generation for Gray/Color Images 基于灰度/彩色图像最优密钥生成的轻量级视觉符号密码方案的最优多秘密图像共享
IF 1.6 Q4 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2023-07-28 DOI: 10.1142/s0219467825500172
Pramod M. Bachiphale, N. Zulpe
Problem: Digital devices are becoming increasingly powerful and smart, which is improving quality of life, but presents new challenges to privacy protection. Visual cryptographic schemes provide data sharing privacy, but have drawbacks such as extra storage space, lossy secret images, and the need to store permutation keys. Aim: This paper proposes a light-weight visual sign-cryptography scheme based on optimal key generation to address the disadvantages of existing visual cryptographic schemes and improve the security, sharing quality, and time consumption of multisecret images. Methods: The proposed light-weight visual sign-cryptography (LW-VSC) scheme consists of three processes: band separation, shares generation, and signcryption/un-signcryption. The process of separation and shares generation is done by an existing method. The multiple shares of the secret images are then encrypted/decrypted using light-weight sign-cryptography. The proposed scheme uses a novel harpy eagle search optimization (HESO) algorithm to generate optimal keys for both the encrypt/decrypt processes. Results: Simulation results and comparative analysis showed the proposed scheme is more secure and requires less storage space, with faster encryption/decryption and improved key generation quality. Conclusion: The proposed light-weight visual sign-cryptography scheme based on optimal key generation is a promising approach to enhance security and improve data sharing quality. The HESO algorithm shows promise in improving the quality of key generation, providing better privacy protection in the face of increasingly powerful digital devices.
问题:数字设备变得越来越强大和智能,这提高了生活质量,但对隐私保护提出了新的挑战。可视化加密方案提供数据共享隐私,但也有缺点,例如额外的存储空间、有损的秘密图像以及需要存储排列密钥。目的:针对现有视觉密码方案的不足,提出了一种基于最优密钥生成的轻量级视觉符号密码方案,提高了多秘密图像的安全性、共享质量和时间消耗。方法:提出了一种轻量级可视签名密码(LW-VSC)方案,该方案包括三个过程:频带分离、共享生成和签名加密/反签名加密。分离和股份生成过程由现有方法完成。然后使用轻量级符号加密技术对秘密图像的多个共享进行加密/解密。该方案采用一种新颖的鹰搜索优化算法(HESO)为加密/解密过程生成最优密钥。结果:仿真结果和对比分析表明,该方案安全性更高,所需存储空间更少,加解密速度更快,密钥生成质量得到提高。结论:提出的基于最优密钥生成的轻量级视觉符号加密方案是一种很有前途的增强安全性和提高数据共享质量的方法。HESO算法有望提高密钥生成的质量,在面对日益强大的数字设备时提供更好的隐私保护。
{"title":"Optimal Multisecret Image Sharing Using Lightweight Visual Sign-Cryptography Scheme With Optimal Key Generation for Gray/Color Images","authors":"Pramod M. Bachiphale, N. Zulpe","doi":"10.1142/s0219467825500172","DOIUrl":"https://doi.org/10.1142/s0219467825500172","url":null,"abstract":"Problem: Digital devices are becoming increasingly powerful and smart, which is improving quality of life, but presents new challenges to privacy protection. Visual cryptographic schemes provide data sharing privacy, but have drawbacks such as extra storage space, lossy secret images, and the need to store permutation keys. Aim: This paper proposes a light-weight visual sign-cryptography scheme based on optimal key generation to address the disadvantages of existing visual cryptographic schemes and improve the security, sharing quality, and time consumption of multisecret images. Methods: The proposed light-weight visual sign-cryptography (LW-VSC) scheme consists of three processes: band separation, shares generation, and signcryption/un-signcryption. The process of separation and shares generation is done by an existing method. The multiple shares of the secret images are then encrypted/decrypted using light-weight sign-cryptography. The proposed scheme uses a novel harpy eagle search optimization (HESO) algorithm to generate optimal keys for both the encrypt/decrypt processes. Results: Simulation results and comparative analysis showed the proposed scheme is more secure and requires less storage space, with faster encryption/decryption and improved key generation quality. Conclusion: The proposed light-weight visual sign-cryptography scheme based on optimal key generation is a promising approach to enhance security and improve data sharing quality. The HESO algorithm shows promise in improving the quality of key generation, providing better privacy protection in the face of increasingly powerful digital devices.","PeriodicalId":44688,"journal":{"name":"International Journal of Image and Graphics","volume":" ","pages":""},"PeriodicalIF":1.6,"publicationDate":"2023-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45471841","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive Total-Variation and Nonconvex Low-Rank Model for Image Denoising 自适应全变分和非凸低秩图像去噪模型
IF 1.6 Q4 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2023-07-27 DOI: 10.1142/s0219467825500160
Li Fang, Wang Xianghai
In recent years, image denoising methods based on total variational regularization have attracted extensive attention. However, the traditional total variational regularization method is an approximate solution based on convex method, and does not consider the particularity of the region with rich details. In this paper, the adaptive total-variation and nonconvex low-rank model for image denoising is proposed, which is a hybrid regularization model. First, the image is decomposed into sparse terms and low rank terms, and then the total variational regularization is used to denoise. At the same time, an adaptive coefficient based on gradient is constructed to adaptively judge the flat area and detail texture area, slow down the denoising intensity of detail area, and then play the role of preserving detail information. Finally, by constructing a nonconvex function, the optimal solution of the function is obtained by using the alternating minimization method. This method not only effectively removes the image noise, but also retains the detailed information of the image. The experimental results show the effectiveness of the proposed model, and SNR and SSIM of the denoised image are improved.
近年来,基于全变分正则化的图像去噪方法引起了人们的广泛关注。然而,传统的全变分正则化方法是基于凸方法的近似解,没有考虑细节丰富区域的特殊性。本文提出了一种用于图像去噪的自适应全变分非凸低秩模型,这是一种混合正则化模型。首先,将图像分解为稀疏项和低秩项,然后使用全变分正则化进行去噪。同时,构造了一个基于梯度的自适应系数,自适应地判断平面区域和细节纹理区域,减缓细节区域的去噪强度,进而起到保存细节信息的作用。最后,通过构造一个非凸函数,利用交替最小化方法得到该函数的最优解。该方法不仅有效地去除了图像噪声,而且保留了图像的详细信息。实验结果表明了该模型的有效性,并且提高了去噪图像的信噪比和SSIM。
{"title":"Adaptive Total-Variation and Nonconvex Low-Rank Model for Image Denoising","authors":"Li Fang, Wang Xianghai","doi":"10.1142/s0219467825500160","DOIUrl":"https://doi.org/10.1142/s0219467825500160","url":null,"abstract":"In recent years, image denoising methods based on total variational regularization have attracted extensive attention. However, the traditional total variational regularization method is an approximate solution based on convex method, and does not consider the particularity of the region with rich details. In this paper, the adaptive total-variation and nonconvex low-rank model for image denoising is proposed, which is a hybrid regularization model. First, the image is decomposed into sparse terms and low rank terms, and then the total variational regularization is used to denoise. At the same time, an adaptive coefficient based on gradient is constructed to adaptively judge the flat area and detail texture area, slow down the denoising intensity of detail area, and then play the role of preserving detail information. Finally, by constructing a nonconvex function, the optimal solution of the function is obtained by using the alternating minimization method. This method not only effectively removes the image noise, but also retains the detailed information of the image. The experimental results show the effectiveness of the proposed model, and SNR and SSIM of the denoised image are improved.","PeriodicalId":44688,"journal":{"name":"International Journal of Image and Graphics","volume":" ","pages":""},"PeriodicalIF":1.6,"publicationDate":"2023-07-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44839070","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of Image and Graphics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1