Pub Date : 2024-09-20eCollection Date: 2024-01-01DOI: 10.7717/peerj-cs.2281
Yanrui Dong
The evolution of social intelligence has led to the adoption of intelligent accounting practices in enterprises. To enhance the efficiency of enterprise accounting operations and improve the capabilities of accountants, we propose an intelligent accounting optimization approach that integrates meta-heuristic algorithms with convolutional neural networks (CNN). First, we enhance the CNN framework by incorporating document and voucher information into accounting audits, creating a multi-modal feature extraction mechanism. Utilizing these multi-modal accounting features, we then introduce a method for assessing accounting quality, which objectively evaluates financial performance. Finally, we propose an optimization technique based on meta-heuristic principles, combining genetic algorithms with annealing models to improve the accounting system. Experimental results validate our approach, demonstrating an accuracy of 0.943 and a mean average precision (mAP) score of 0.812. This method provides technological support for refining accounting audit mechanisms.
{"title":"Intelligent accounting optimization method based on meta-heuristic algorithm and CNN.","authors":"Yanrui Dong","doi":"10.7717/peerj-cs.2281","DOIUrl":"10.7717/peerj-cs.2281","url":null,"abstract":"<p><p>The evolution of social intelligence has led to the adoption of intelligent accounting practices in enterprises. To enhance the efficiency of enterprise accounting operations and improve the capabilities of accountants, we propose an intelligent accounting optimization approach that integrates meta-heuristic algorithms with convolutional neural networks (CNN). First, we enhance the CNN framework by incorporating document and voucher information into accounting audits, creating a multi-modal feature extraction mechanism. Utilizing these multi-modal accounting features, we then introduce a method for assessing accounting quality, which objectively evaluates financial performance. Finally, we propose an optimization technique based on meta-heuristic principles, combining genetic algorithms with annealing models to improve the accounting system. Experimental results validate our approach, demonstrating an accuracy of 0.943 and a mean average precision (mAP) score of 0.812. This method provides technological support for refining accounting audit mechanisms.</p>","PeriodicalId":54224,"journal":{"name":"PeerJ Computer Science","volume":"10 ","pages":"e2281"},"PeriodicalIF":3.5,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11419630/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142309136","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-20eCollection Date: 2024-01-01DOI: 10.7717/peerj-cs.2181
Rabbia Zia, Mariam Rehman, Afzaal Hussain, Shahbaz Nazeer, Maria Anjum
Synthetic images ar---e created using computer graphics modeling and artificial intelligence techniques, referred to as deepfakes. They modify human features by using generative models and deep learning algorithms, posing risks violations of social media regulations and spread false information. To address these concerns, the study proposed an improved generative adversarial network (GAN) model which improves accuracy while differentiating between real and fake images focusing on data augmentation and label smoothing strategies for GAN training. The study utilizes a dataset containing human faces and employs DCGAN (deep convolutional generative adversarial network) as the base model. In comparison with the traditional GANs, the proposed GAN outperform in terms of frequently used metrics i.e., Fréchet Inception Distance (FID) and accuracy. The model effectiveness is demonstrated through evaluation on the Flickr-Faces Nvidia dataset and Fakefaces d--ataset, achieving an FID score of 55.67, an accuracy of 98.82%, and an F1-score of 0.99 in detection. This study optimizes the model parameters to achieve optimal parameter settings. This study fine-tune the model parameters to reach optimal settings, thereby reducing risks in synthetic image generation. The article introduces an effective framework for both image manipulation and detection.
利用计算机图形建模和人工智能技术创建的合成图像被称为深度伪造。它们利用生成模型和深度学习算法修改人体特征,有可能违反社交媒体法规并传播虚假信息。为了解决这些问题,该研究提出了一种改进的生成对抗网络(GAN)模型,该模型在提高准确性的同时,还能区分真实和虚假图像,重点关注 GAN 训练中的数据增强和标签平滑策略。该研究利用包含人脸的数据集,并采用 DCGAN(深度卷积生成式对抗网络)作为基础模型。与传统的 GAN 相比,所提出的 GAN 在常用指标(即弗雷谢特起始距离(FID)和准确率)方面表现更优。通过对 Flickr-Faces Nvidia 数据集和 Fakefaces d--ataset 数据集的评估,证明了该模型的有效性,其 FID 得分为 55.67,准确率为 98.82%,检测的 F1 分数为 0.99。本研究对模型参数进行了优化,以达到最佳参数设置。本研究对模型参数进行微调,以达到最佳参数设置,从而降低合成图像生成的风险。文章介绍了一种有效的图像处理和检测框架。
{"title":"Improving synthetic media generation and detection using generative adversarial networks.","authors":"Rabbia Zia, Mariam Rehman, Afzaal Hussain, Shahbaz Nazeer, Maria Anjum","doi":"10.7717/peerj-cs.2181","DOIUrl":"10.7717/peerj-cs.2181","url":null,"abstract":"<p><p>Synthetic images ar---e created using computer graphics modeling and artificial intelligence techniques, referred to as deepfakes. They modify human features by using generative models and deep learning algorithms, posing risks violations of social media regulations and spread false information. To address these concerns, the study proposed an improved generative adversarial network (GAN) model which improves accuracy while differentiating between real and fake images focusing on data augmentation and label smoothing strategies for GAN training. The study utilizes a dataset containing human faces and employs DCGAN (deep convolutional generative adversarial network) as the base model. In comparison with the traditional GANs, the proposed GAN outperform in terms of frequently used metrics <i>i.e</i>., Fréchet Inception Distance (FID) and accuracy. The model effectiveness is demonstrated through evaluation on the Flickr-Faces Nvidia dataset and Fakefaces d--ataset, achieving an FID score of 55.67, an accuracy of 98.82%, and an F1-score of 0.99 in detection. This study optimizes the model parameters to achieve optimal parameter settings. This study fine-tune the model parameters to reach optimal settings, thereby reducing risks in synthetic image generation. The article introduces an effective framework for both image manipulation and detection.</p>","PeriodicalId":54224,"journal":{"name":"PeerJ Computer Science","volume":"10 ","pages":"e2181"},"PeriodicalIF":3.5,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11419665/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142309135","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-20eCollection Date: 2024-01-01DOI: 10.7717/peerj-cs.2252
Mohammed Abdalsalam, Chunlin Li, Abdelghani Dahou, Natalia Kryvinska
The world faces the ongoing challenge of terrorism and extremism, which threaten the stability of nations, the security of their citizens, and the integrity of political, economic, and social systems. Given the complexity and multifaceted nature of this phenomenon, combating it requires a collective effort, with tailored methods to address its various aspects. Identifying the terrorist organization responsible for an attack is a critical step in combating terrorism. Historical data plays a pivotal role in this process, providing insights that can inform prevention and response strategies. With advancements in technology and artificial intelligence (AI), particularly in military applications, there is growing interest in utilizing these developments to enhance national and regional security against terrorism. Central to this effort are terrorism databases, which serve as rich resources for data on armed organizations, extremist entities, and terrorist incidents. The Global Terrorism Database (GTD) stands out as one of the most widely used and accessible resources for researchers. Recent progress in machine learning (ML), deep learning (DL), and natural language processing (NLP) offers promising avenues for improving the identification and classification of terrorist organizations. This study introduces a framework designed to classify and predict terrorist groups using bidirectional recurrent units and self-attention mechanisms, referred to as BiGRU-SA. This approach utilizes the comprehensive data in the GTD by integrating textual features extracted by DistilBERT with features that show a high correlation with terrorist organizations. Additionally, the Synthetic Minority Over-sampling Technique with Tomek links (SMOTE-T) was employed to address data imbalance and enhance the robustness of our predictions. The BiGRU-SA model captures temporal dependencies and contextual information within the data. By processing data sequences in both forward and reverse directions, BiGRU-SA offers a comprehensive view of the temporal dynamics, significantly enhancing classification accuracy. To evaluate the effectiveness of our framework, we compared ten models, including six traditional ML models and four DL algorithms. The proposed BiGRU-SA framework demonstrated outstanding performance in classifying 36 terrorist organizations responsible for terrorist attacks, achieving an accuracy of 98.68%, precision of 96.06%, sensitivity of 96.83%, specificity of 99.50%, and a Matthews correlation coefficient of 97.50%. Compared to state-of-the-art methods, the proposed model outperformed others, confirming its effectiveness and accuracy in the classification and prediction of terrorist organizations.
{"title":"Terrorism group prediction using feature combination and BiGRU with self-attention mechanism.","authors":"Mohammed Abdalsalam, Chunlin Li, Abdelghani Dahou, Natalia Kryvinska","doi":"10.7717/peerj-cs.2252","DOIUrl":"10.7717/peerj-cs.2252","url":null,"abstract":"<p><p>The world faces the ongoing challenge of terrorism and extremism, which threaten the stability of nations, the security of their citizens, and the integrity of political, economic, and social systems. Given the complexity and multifaceted nature of this phenomenon, combating it requires a collective effort, with tailored methods to address its various aspects. Identifying the terrorist organization responsible for an attack is a critical step in combating terrorism. Historical data plays a pivotal role in this process, providing insights that can inform prevention and response strategies. With advancements in technology and artificial intelligence (AI), particularly in military applications, there is growing interest in utilizing these developments to enhance national and regional security against terrorism. Central to this effort are terrorism databases, which serve as rich resources for data on armed organizations, extremist entities, and terrorist incidents. The Global Terrorism Database (GTD) stands out as one of the most widely used and accessible resources for researchers. Recent progress in machine learning (ML), deep learning (DL), and natural language processing (NLP) offers promising avenues for improving the identification and classification of terrorist organizations. This study introduces a framework designed to classify and predict terrorist groups using bidirectional recurrent units and self-attention mechanisms, referred to as BiGRU-SA. This approach utilizes the comprehensive data in the GTD by integrating textual features extracted by DistilBERT with features that show a high correlation with terrorist organizations. Additionally, the Synthetic Minority Over-sampling Technique with Tomek links (SMOTE-T) was employed to address data imbalance and enhance the robustness of our predictions. The BiGRU-SA model captures temporal dependencies and contextual information within the data. By processing data sequences in both forward and reverse directions, BiGRU-SA offers a comprehensive view of the temporal dynamics, significantly enhancing classification accuracy. To evaluate the effectiveness of our framework, we compared ten models, including six traditional ML models and four DL algorithms. The proposed BiGRU-SA framework demonstrated outstanding performance in classifying 36 terrorist organizations responsible for terrorist attacks, achieving an accuracy of 98.68%, precision of 96.06%, sensitivity of 96.83%, specificity of 99.50%, and a Matthews correlation coefficient of 97.50%. Compared to state-of-the-art methods, the proposed model outperformed others, confirming its effectiveness and accuracy in the classification and prediction of terrorist organizations.</p>","PeriodicalId":54224,"journal":{"name":"PeerJ Computer Science","volume":"10 ","pages":"e2252"},"PeriodicalIF":3.5,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11419613/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142309137","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-20eCollection Date: 2024-01-01DOI: 10.7717/peerj-cs.2264
Muhammad Rehan Naeem, Rashid Amin, Muhammad Farhan, Faiz Abdullah Alotaibi, Mrim M Alnfiai, Gabriel Avelino Sampedro, Vincent Karovič
Collective intelligence systems like Chat Generative Pre-Trained Transformer (ChatGPT) have emerged. They have brought both promise and peril to cybersecurity and privacy protection. This study introduces novel approaches to harness the power of artificial intelligence (AI) and big data analytics to enhance security and privacy in this new era. Contributions could explore topics such as: leveraging natural language processing (NLP) in ChatGPT-like systems to strengthen information security; evaluating privacy-enhancing technologies to maximize data utility while minimizing personal data exposure; modeling human behavior and agency to build secure and ethical human-centric systems; applying machine learning to detect threats and vulnerabilities in a data-driven manner; using analytics to preserve privacy in large datasets while enabling value creation; crafting AI techniques that operate in a trustworthy and explainable manner. This article advances the state-of-the-art at the intersection of cybersecurity, privacy, human factors, ethics, and cutting-edge AI, providing impactful solutions to emerging challenges. Our research presents a revolutionary approach to malware detection that leverages deep learning (DL) based methodologies to automatically learn features from raw data. Our approach involves constructing a grayscale image from a malware file and extracting features to minimize its size. This process affords us the ability to discern patterns that might remain hidden from other techniques, enabling us to utilize convolutional neural networks (CNNs) to learn from these grayscale images and a stacking ensemble to classify malware. The goal is to model a highly complex nonlinear function with parameters that can be optimized to achieve superior performance. To test our approach, we ran it on over 6,414 malware variants and 2,050 benign files from the MalImg collection, resulting in an impressive 99.86 percent validation accuracy for malware detection. Furthermore, we conducted a classification experiment on 15 malware families and 13 tests with varying parameters to compare our model to other comparable research. Our model outperformed most of the similar research with detection accuracy ranging from 47.07% to 99.81% and a significant increase in detection performance. Our results demonstrate the efficacy of our approach, which unlocks the hidden patterns that underlie complex systems, advancing the frontiers of computational security.
{"title":"Harnessing AI and analytics to enhance cybersecurity and privacy for collective intelligence systems.","authors":"Muhammad Rehan Naeem, Rashid Amin, Muhammad Farhan, Faiz Abdullah Alotaibi, Mrim M Alnfiai, Gabriel Avelino Sampedro, Vincent Karovič","doi":"10.7717/peerj-cs.2264","DOIUrl":"10.7717/peerj-cs.2264","url":null,"abstract":"<p><p>Collective intelligence systems like Chat Generative Pre-Trained Transformer (ChatGPT) have emerged. They have brought both promise and peril to cybersecurity and privacy protection. This study introduces novel approaches to harness the power of artificial intelligence (AI) and big data analytics to enhance security and privacy in this new era. Contributions could explore topics such as: leveraging natural language processing (NLP) in ChatGPT-like systems to strengthen information security; evaluating privacy-enhancing technologies to maximize data utility while minimizing personal data exposure; modeling human behavior and agency to build secure and ethical human-centric systems; applying machine learning to detect threats and vulnerabilities in a data-driven manner; using analytics to preserve privacy in large datasets while enabling value creation; crafting AI techniques that operate in a trustworthy and explainable manner. This article advances the state-of-the-art at the intersection of cybersecurity, privacy, human factors, ethics, and cutting-edge AI, providing impactful solutions to emerging challenges. Our research presents a revolutionary approach to malware detection that leverages deep learning (DL) based methodologies to automatically learn features from raw data. Our approach involves constructing a grayscale image from a malware file and extracting features to minimize its size. This process affords us the ability to discern patterns that might remain hidden from other techniques, enabling us to utilize convolutional neural networks (CNNs) to learn from these grayscale images and a stacking ensemble to classify malware. The goal is to model a highly complex nonlinear function with parameters that can be optimized to achieve superior performance. To test our approach, we ran it on over 6,414 malware variants and 2,050 benign files from the MalImg collection, resulting in an impressive 99.86 percent validation accuracy for malware detection. Furthermore, we conducted a classification experiment on 15 malware families and 13 tests with varying parameters to compare our model to other comparable research. Our model outperformed most of the similar research with detection accuracy ranging from 47.07% to 99.81% and a significant increase in detection performance. Our results demonstrate the efficacy of our approach, which unlocks the hidden patterns that underlie complex systems, advancing the frontiers of computational security.</p>","PeriodicalId":54224,"journal":{"name":"PeerJ Computer Science","volume":"10 ","pages":"e2264"},"PeriodicalIF":3.5,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11419604/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142309134","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-20eCollection Date: 2024-01-01DOI: 10.7717/peerj-cs.2202
Drishti Sharma, Abhishek Srivastava
Social media, an undeniable facet of the modern era, has become a primary pathway for disseminating information. Unverified and potentially harmful rumors can have detrimental effects on both society and individuals. Owing to the plethora of content generated, it is essential to assess its alignment with factual accuracy and determine its veracity. Previous research has explored various approaches, including feature engineering and deep learning techniques, that leverage propagation theory to identify rumors. In our study, we place significant importance on examining the emotional and sentimental aspects of tweets using deep learning approaches to improve our ability to detect rumors. Leveraging the findings from the previous analysis, we propose a Sentiment and EMotion driven TransformEr Classifier method (SEMTEC). Unlike the existing studies, our method leverages the extraction of emotion and sentiment tags alongside the assimilation of the content-based information from the textual modality, i.e., the main tweet. This meticulous semantic analysis allows us to measure the user's emotional state, leading to an impressive accuracy rate of 92% for rumor detection on the "PHEME" dataset. The validation is carried out on a novel dataset named "Twitter24". Furthermore, SEMTEC exceeds standard methods accuracy by around 2% on "Twitter24" dataset.
{"title":"Detecting rumors in social media using emotion based deep learning approach.","authors":"Drishti Sharma, Abhishek Srivastava","doi":"10.7717/peerj-cs.2202","DOIUrl":"10.7717/peerj-cs.2202","url":null,"abstract":"<p><p>Social media, an undeniable facet of the modern era, has become a primary pathway for disseminating information. Unverified and potentially harmful rumors can have detrimental effects on both society and individuals. Owing to the plethora of content generated, it is essential to assess its alignment with factual accuracy and determine its veracity. Previous research has explored various approaches, including feature engineering and deep learning techniques, that leverage propagation theory to identify rumors. In our study, we place significant importance on examining the emotional and sentimental aspects of tweets using deep learning approaches to improve our ability to detect rumors. Leveraging the findings from the previous analysis, we propose a Sentiment and EMotion driven TransformEr Classifier method (SEMTEC). Unlike the existing studies, our method leverages the extraction of emotion and sentiment tags alongside the assimilation of the content-based information from the textual modality, <i>i.e.</i>, the main tweet. This meticulous semantic analysis allows us to measure the user's emotional state, leading to an impressive accuracy rate of 92% for rumor detection on the \"PHEME\" dataset. The validation is carried out on a novel dataset named \"Twitter24\". Furthermore, SEMTEC exceeds standard methods accuracy by around 2% on \"Twitter24\" dataset.</p>","PeriodicalId":54224,"journal":{"name":"PeerJ Computer Science","volume":"10 ","pages":"e2202"},"PeriodicalIF":3.5,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11419649/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142309133","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Image style transfer is an important way to combine different styles and contents to generate new images, which plays an important role in computer vision tasks such as image reconstruction and image texture synthesis. In style transfer tasks, there are often long-distance dependencies between pixels of different styles and contents, and existing neural network-based work cannot handle this problem well. This paper constructs a generation model for style transfer based on the cycle-consistent network and the attention mechanism. The forward and backward learning process of the cycle-consistent mechanism could make the network complete the mismatch conversion between the input and output of the image. The attention mechanism enhances the model's ability to perceive the long-distance dependencies between pixels in process of learning feature representation from the target content and the target styles, and at the same time suppresses the style feature information of the non-target area. Finally, a large number of experiments were carried out in the monet2photo dataset, and the results show that the misjudgment rate of Amazon Mechanical Turk (AMT) perceptual studies achieves 45%, which verified that the cycle-consistent network model with attention mechanism has certain advantages in image style transfer.
{"title":"A model integrating attention mechanism and generative adversarial network for image style transfer.","authors":"Miaomiao Fu, Yixing Liu, Rongrong Ma, Binbin Zhang, Linli Wu, Lingli Zhu","doi":"10.7717/peerj-cs.2332","DOIUrl":"10.7717/peerj-cs.2332","url":null,"abstract":"<p><p>Image style transfer is an important way to combine different styles and contents to generate new images, which plays an important role in computer vision tasks such as image reconstruction and image texture synthesis. In style transfer tasks, there are often long-distance dependencies between pixels of different styles and contents, and existing neural network-based work cannot handle this problem well. This paper constructs a generation model for style transfer based on the cycle-consistent network and the attention mechanism. The forward and backward learning process of the cycle-consistent mechanism could make the network complete the mismatch conversion between the input and output of the image. The attention mechanism enhances the model's ability to perceive the long-distance dependencies between pixels in process of learning feature representation from the target content and the target styles, and at the same time suppresses the style feature information of the non-target area. Finally, a large number of experiments were carried out in the monet2photo dataset, and the results show that the misjudgment rate of Amazon Mechanical Turk (AMT) perceptual studies achieves 45%, which verified that the cycle-consistent network model with attention mechanism has certain advantages in image style transfer.</p>","PeriodicalId":54224,"journal":{"name":"PeerJ Computer Science","volume":"10 ","pages":"e2332"},"PeriodicalIF":3.5,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11419672/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142309132","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Early fire warning is critical to the safety and stability of power systems. However, current methods encounter challenges in capturing subtle features, limiting their effectiveness in providing timely alerts for potential fire hazards. To overcome this drawback, a novel detection algorithm for thin smoke was proposed to enhance early fire detection capabilities. The core is that the Patch-TBV feature was proposed first, and the total bounded variation (TBV) was computed at the patch level. This approach is rooted in the understanding that traditional methods struggle to detect minute variations in image characteristics, particularly in scenarios where the features are dispersed or subtle. By computing TBV at a more localized level, the algorithm proposed gains a finer granularity in assessing image quality, enabling it to capture subtle variations that might indicate the presence of smoke or early signs of a fire. Another key aspect that sets our algorithm apart is the incorporation of subtle variation magnification. This technique serves to magnify subtle features within the image, leveraging the computed TBV values. This magnification strategy is pivotal for improving the algorithm’s precision in detecting subtle variations, especially in environments where smoke concentrations may be minimal or dispersed. To evaluate the algorithm’s performance in real-world scenarios, a comprehensive dataset, named TIP, comprising 3,120 images was constructed. The dataset covers diverse conditions and potential challenges that might be encountered in practical applications. Experimental results confirm the robustness and effectiveness of the proposed algorithm, showcasing its ability to provide accurate and timely fire warnings in various contexts. In conclusion, our research not only identifies the limitations of existing methods in capturing subtle features for early fire detection but also proposes a sophisticated algorithm, integrating Patch-TBV and micro-variation amplification, to address these challenges. The algorithm’s effectiveness and robustness are substantiated through extensive testing, demonstrating its potential as a valuable tool for enhancing fire safety in power systems and similar environments.
{"title":"A flexible perception method of thin smoke based on patch total bounded variation for buildings","authors":"Jieming Zhang, Yifan Gao, Xianchao Chen, Zhanchen Chen","doi":"10.7717/peerj-cs.2282","DOIUrl":"https://doi.org/10.7717/peerj-cs.2282","url":null,"abstract":"Early fire warning is critical to the safety and stability of power systems. However, current methods encounter challenges in capturing subtle features, limiting their effectiveness in providing timely alerts for potential fire hazards. To overcome this drawback, a novel detection algorithm for thin smoke was proposed to enhance early fire detection capabilities. The core is that the Patch-TBV feature was proposed first, and the total bounded variation (TBV) was computed at the patch level. This approach is rooted in the understanding that traditional methods struggle to detect minute variations in image characteristics, particularly in scenarios where the features are dispersed or subtle. By computing TBV at a more localized level, the algorithm proposed gains a finer granularity in assessing image quality, enabling it to capture subtle variations that might indicate the presence of smoke or early signs of a fire. Another key aspect that sets our algorithm apart is the incorporation of subtle variation magnification. This technique serves to magnify subtle features within the image, leveraging the computed TBV values. This magnification strategy is pivotal for improving the algorithm’s precision in detecting subtle variations, especially in environments where smoke concentrations may be minimal or dispersed. To evaluate the algorithm’s performance in real-world scenarios, a comprehensive dataset, named TIP, comprising 3,120 images was constructed. The dataset covers diverse conditions and potential challenges that might be encountered in practical applications. Experimental results confirm the robustness and effectiveness of the proposed algorithm, showcasing its ability to provide accurate and timely fire warnings in various contexts. In conclusion, our research not only identifies the limitations of existing methods in capturing subtle features for early fire detection but also proposes a sophisticated algorithm, integrating Patch-TBV and micro-variation amplification, to address these challenges. The algorithm’s effectiveness and robustness are substantiated through extensive testing, demonstrating its potential as a valuable tool for enhancing fire safety in power systems and similar environments.","PeriodicalId":54224,"journal":{"name":"PeerJ Computer Science","volume":"46 1","pages":""},"PeriodicalIF":3.8,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142247784","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Noemi Merayo, Alba Ayuso-Lanchares, Clara González-Sanguino
Background This study aims to examine, through artificial intelligence, specifically machine learning, the emotional impact generated by disclosures about mental health on social media. In contrast to previous research, which primarily focused on identifying psychopathologies, our study investigates the emotional response to mental health-related content on Instagram, particularly content created by influencers/celebrities. This platform, especially favored by the youth, is the stage where these influencers exert significant social impact, and where their analysis holds strong relevance. Analyzing mental health with machine learning techniques on Instagram is unprecedented, as all existing research has primarily focused on Twitter. Methods This research involves creating a new corpus labelled with responses to mental health posts made by influencers/celebrities on Instagram, categorized by emotions such as love/admiration, anger/contempt/mockery, gratitude, identification/empathy, and sadness. The study is complemented by modelling a set of machine learning algorithms to efficiently detect the emotions arising when faced with these mental health disclosures on Instagram, using the previous corpus. Results Results have shown that machine learning algorithms can effectively detect such emotional responses. Traditional techniques, such as Random Forest, showed decent performance with low computational loads (around 50%), while deep learning and Bidirectional Encoder Representation from Transformers (BERT) algorithms achieved very good results. In particular, the BERT models reached accuracy levels between 86–90%, and the deep learning model achieved 72% accuracy. These results are satisfactory, considering that predicting emotions, especially in social networks, is challenging due to factors such as the subjectivity of emotion interpretation, the variability of emotions between individuals, and the interpretation of emotions in different cultures and communities. Discussion This cross-cutting research between mental health and artificial intelligence allows us to understand the emotional impact generated by mental health content on social networks, especially content generated by influential celebrities among young people. The application of machine learning allows us to understand the emotional reactions of society to messages related to mental health, which is highly innovative and socially relevant given the importance of the phenomenon in societies. In fact, the proposed algorithms’ high accuracy (86–90%) in social contexts like mental health, where detecting negative emotions is crucial, presents a promising research avenue. Achieving such levels of accuracy is highly valuable due to the significant implications of false positives or false negatives in this social context.
{"title":"Machine learning and natural language processing to assess the emotional impact of influencers’ mental health content on Instagram","authors":"Noemi Merayo, Alba Ayuso-Lanchares, Clara González-Sanguino","doi":"10.7717/peerj-cs.2251","DOIUrl":"https://doi.org/10.7717/peerj-cs.2251","url":null,"abstract":"Background\u0000This study aims to examine, through artificial intelligence, specifically machine learning, the emotional impact generated by disclosures about mental health on social media. In contrast to previous research, which primarily focused on identifying psychopathologies, our study investigates the emotional response to mental health-related content on Instagram, particularly content created by influencers/celebrities. This platform, especially favored by the youth, is the stage where these influencers exert significant social impact, and where their analysis holds strong relevance. Analyzing mental health with machine learning techniques on Instagram is unprecedented, as all existing research has primarily focused on Twitter. Methods\u0000This research involves creating a new corpus labelled with responses to mental health posts made by influencers/celebrities on Instagram, categorized by emotions such as love/admiration, anger/contempt/mockery, gratitude, identification/empathy, and sadness. The study is complemented by modelling a set of machine learning algorithms to efficiently detect the emotions arising when faced with these mental health disclosures on Instagram, using the previous corpus. Results\u0000Results have shown that machine learning algorithms can effectively detect such emotional responses. Traditional techniques, such as Random Forest, showed decent performance with low computational loads (around 50%), while deep learning and Bidirectional Encoder Representation from Transformers (BERT) algorithms achieved very good results. In particular, the BERT models reached accuracy levels between 86–90%, and the deep learning model achieved 72% accuracy. These results are satisfactory, considering that predicting emotions, especially in social networks, is challenging due to factors such as the subjectivity of emotion interpretation, the variability of emotions between individuals, and the interpretation of emotions in different cultures and communities. Discussion\u0000This cross-cutting research between mental health and artificial intelligence allows us to understand the emotional impact generated by mental health content on social networks, especially content generated by influential celebrities among young people. The application of machine learning allows us to understand the emotional reactions of society to messages related to mental health, which is highly innovative and socially relevant given the importance of the phenomenon in societies. In fact, the proposed algorithms’ high accuracy (86–90%) in social contexts like mental health, where detecting negative emotions is crucial, presents a promising research avenue. Achieving such levels of accuracy is highly valuable due to the significant implications of false positives or false negatives in this social context.","PeriodicalId":54224,"journal":{"name":"PeerJ Computer Science","volume":"26 1","pages":""},"PeriodicalIF":3.8,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142268378","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Hafsa Dar, Romana Aziz, Javed Ali Khan, Muhammad IkramUllah Lali, Nouf Abdullah Almujally
Ambiguity is a common challenge in specifying natural language (NL) requirements. One of the reasons for the occurrence of ambiguity in software requirements is the lack of user involvement in requirements elicitation and inspection phases. Even if they get involved, it is hard for them to understand the context of the system, and ultimately unable to provide requirements correctly due to a lack of interest. Previously, the researchers have worked on ambiguity avoidance, detection, and removal techniques in requirements. Still, less work is reported in the literature to actively engage users in the system to reduce ambiguity at the early stages of requirements engineering. Traditionally, ambiguity is addressed during inspection when requirements are initially specified in the SRS document. Resolving or removing ambiguity during the inspection is time-consuming, costly, and laborious. Also, traditional elicitation techniques have limitations like lack of user involvement, inactive user participation, biases, incomplete requirements, etc. Therefore, in this study, we have designed a framework, Gamification for Lexical Ambiguity (Gamify4LexAmb), for detecting and reducing ambiguity using gamification. Gamify4LexAmb engages users and identifies lexical ambiguity in requirements, which occurs in polysemy words where a single word can have several different meanings. We have also validated Gamify4LexAmb by developing an initial prototype. The results show that Gamify4LexAmb successfully identifies lexical ambiguities in given requirements by engaging users in requirements elicitation. In the next part of our research, an industrial case study will be performed to understand the effects of gamification on real-time data for detecting and reducing NL ambiguity.
{"title":"Gamify4LexAmb: a gamification-based approach to address lexical ambiguity in natural language requirements","authors":"Hafsa Dar, Romana Aziz, Javed Ali Khan, Muhammad IkramUllah Lali, Nouf Abdullah Almujally","doi":"10.7717/peerj-cs.2229","DOIUrl":"https://doi.org/10.7717/peerj-cs.2229","url":null,"abstract":"Ambiguity is a common challenge in specifying natural language (NL) requirements. One of the reasons for the occurrence of ambiguity in software requirements is the lack of user involvement in requirements elicitation and inspection phases. Even if they get involved, it is hard for them to understand the context of the system, and ultimately unable to provide requirements correctly due to a lack of interest. Previously, the researchers have worked on ambiguity avoidance, detection, and removal techniques in requirements. Still, less work is reported in the literature to actively engage users in the system to reduce ambiguity at the early stages of requirements engineering. Traditionally, ambiguity is addressed during inspection when requirements are initially specified in the SRS document. Resolving or removing ambiguity during the inspection is time-consuming, costly, and laborious. Also, traditional elicitation techniques have limitations like lack of user involvement, inactive user participation, biases, incomplete requirements, etc. Therefore, in this study, we have designed a framework, Gamification for Lexical Ambiguity (Gamify4LexAmb), for detecting and reducing ambiguity using gamification. Gamify4LexAmb engages users and identifies lexical ambiguity in requirements, which occurs in polysemy words where a single word can have several different meanings. We have also validated Gamify4LexAmb by developing an initial prototype. The results show that Gamify4LexAmb successfully identifies lexical ambiguities in given requirements by engaging users in requirements elicitation. In the next part of our research, an industrial case study will be performed to understand the effects of gamification on real-time data for detecting and reducing NL ambiguity.","PeriodicalId":54224,"journal":{"name":"PeerJ Computer Science","volume":"212 1","pages":""},"PeriodicalIF":3.8,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142268379","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Background Early diagnosis and treatment of diabetic eye disease (DED) improve prognosis and lessen the possibility of permanent vision loss. Screening of retinal fundus images is a significant process widely employed for diagnosing patients with DED or other eye problems. However, considerable time and effort are required to detect these images manually. Methods Deep learning approaches in machine learning have attained superior performance for the binary classification of healthy and pathological retinal fundus images. In contrast, multi-class retinal eye disease classification is still a difficult task. Therefore, a two-phase transfer learning approach is developed in this research for automated classification and segmentation of multi-class DED pathologies. Results In the first step, a Modified ResNet-50 model pre-trained on the ImageNet dataset was transferred and learned to classify normal diabetic macular edema (DME), diabetic retinopathy, glaucoma, and cataracts. In the second step, the defective region of multiple eye diseases is segmented using the transfer learning-based DenseUNet model. From the publicly accessible dataset, the suggested model is assessed using several retinal fundus images. Our proposed model for multi-class classification achieves a maximum specificity of 99.73%, a sensitivity of 99.54%, and an accuracy of 99.67%.
{"title":"Detection and diagnosis of diabetic eye diseases using two phase transfer learning approach","authors":"Vamsi Krishna Madduri, Battula Srinivasa Rao","doi":"10.7717/peerj-cs.2135","DOIUrl":"https://doi.org/10.7717/peerj-cs.2135","url":null,"abstract":"Background\u0000Early diagnosis and treatment of diabetic eye disease (DED) improve prognosis and lessen the possibility of permanent vision loss. Screening of retinal fundus images is a significant process widely employed for diagnosing patients with DED or other eye problems. However, considerable time and effort are required to detect these images manually. Methods\u0000Deep learning approaches in machine learning have attained superior performance for the binary classification of healthy and pathological retinal fundus images. In contrast, multi-class retinal eye disease classification is still a difficult task. Therefore, a two-phase transfer learning approach is developed in this research for automated classification and segmentation of multi-class DED pathologies. Results\u0000In the first step, a Modified ResNet-50 model pre-trained on the ImageNet dataset was transferred and learned to classify normal diabetic macular edema (DME), diabetic retinopathy, glaucoma, and cataracts. In the second step, the defective region of multiple eye diseases is segmented using the transfer learning-based DenseUNet model. From the publicly accessible dataset, the suggested model is assessed using several retinal fundus images. Our proposed model for multi-class classification achieves a maximum specificity of 99.73%, a sensitivity of 99.54%, and an accuracy of 99.67%.","PeriodicalId":54224,"journal":{"name":"PeerJ Computer Science","volume":"20 1","pages":""},"PeriodicalIF":3.8,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142268380","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}