Pub Date : 2024-10-17DOI: 10.1007/s10462-024-10956-3
Yuxin Ma, Jiaxing Yin, Feng Huang, Qipeng Li
One of the focal points in industrial product defect detection lies in the utilization of deep learning-based object detection algorithms. With the continuous introduction of these algorithms and their refined models, notable achievements have been attained. However, challenges persist in industrial settings, such as substantial variations in defect scales, the delicate balance between accuracy and speed, and the detection of small objects. Various methods have been proposed to address these challenges and propel the advancement of defect detection. To comprehensively review the latest developments in deep learning-based industrial product defect detection algorithms and foster further progress, this paper encompasses typical datasets and evaluation metrics used in industrial product defect detection, traces the development history of supervised one-stage and two-stage object detection algorithm-based and unsupervised algorithm-based industrial defect detection methods, discusses major challenges, and outlines future directions. It highlights the potential for further improving the accuracy, speed, and reliability of defect detection systems in industrial applications.
{"title":"Surface defect inspection of industrial products with object detection deep networks: a systematic review","authors":"Yuxin Ma, Jiaxing Yin, Feng Huang, Qipeng Li","doi":"10.1007/s10462-024-10956-3","DOIUrl":"10.1007/s10462-024-10956-3","url":null,"abstract":"<div><p>One of the focal points in industrial product defect detection lies in the utilization of deep learning-based object detection algorithms. With the continuous introduction of these algorithms and their refined models, notable achievements have been attained. However, challenges persist in industrial settings, such as substantial variations in defect scales, the delicate balance between accuracy and speed, and the detection of small objects. Various methods have been proposed to address these challenges and propel the advancement of defect detection. To comprehensively review the latest developments in deep learning-based industrial product defect detection algorithms and foster further progress, this paper encompasses typical datasets and evaluation metrics used in industrial product defect detection, traces the development history of supervised one-stage and two-stage object detection algorithm-based and unsupervised algorithm-based industrial defect detection methods, discusses major challenges, and outlines future directions. It highlights the potential for further improving the accuracy, speed, and reliability of defect detection systems in industrial applications.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":null,"pages":null},"PeriodicalIF":10.7,"publicationDate":"2024-10-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10956-3.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142443225","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-17DOI: 10.1007/s10462-024-10981-2
Abdelazim G. Hussien, Farhad Soleimanian Gharehchopogh, Anas Bouaouda, Sumit Kumar, Gang Hu
The African Vultures Optimization Algorithm (AVOA) is a recently developed meta-heuristic algorithm inspired by the foraging behavior of African vultures in nature. This algorithm has gained attention due to its simplicity, flexibility, and effectiveness in tackling many optimization problems. The significance of this review lies in its comprehensive examination of the AVOA’s development, core principles, and applications. By analyzing 112 studies, this review highlights the algorithm’s versatility and the growing interest in enhancing its performance for real-world optimization challenges. This review methodically explores the evolution of AVOA, investigating proposed improvements that enhance the algorithm’s ability to adapt to various search geometries in optimization problems. Additionally, it introduces the AVOA solver, detailing its functionality and application in different optimization scenarios. The review demonstrates the AVOA’s effectiveness, particularly its unique weighting mechanism, which mimics vulture behavior during the search process. The findings underscore the algorithm’s robustness, ease of use, and lack of dependence on derivative information. The review also critically evaluates the AVOA’s convergence behavior, identifying its strengths and limitations. In conclusion, the study not only consolidates the existing knowledge on AVOA but also proposes directions for future research, including potential adaptations and enhancements to address its limitations. The insights gained from this review offer valuable guidance for researchers and practitioners seeking to apply or improve the AVOA in various optimization tasks.
{"title":"Recent applications and advances of African Vultures Optimization Algorithm","authors":"Abdelazim G. Hussien, Farhad Soleimanian Gharehchopogh, Anas Bouaouda, Sumit Kumar, Gang Hu","doi":"10.1007/s10462-024-10981-2","DOIUrl":"10.1007/s10462-024-10981-2","url":null,"abstract":"<div><p>The African Vultures Optimization Algorithm (AVOA) is a recently developed meta-heuristic algorithm inspired by the foraging behavior of African vultures in nature. This algorithm has gained attention due to its simplicity, flexibility, and effectiveness in tackling many optimization problems. The significance of this review lies in its comprehensive examination of the AVOA’s development, core principles, and applications. By analyzing 112 studies, this review highlights the algorithm’s versatility and the growing interest in enhancing its performance for real-world optimization challenges. This review methodically explores the evolution of AVOA, investigating proposed improvements that enhance the algorithm’s ability to adapt to various search geometries in optimization problems. Additionally, it introduces the AVOA solver, detailing its functionality and application in different optimization scenarios. The review demonstrates the AVOA’s effectiveness, particularly its unique weighting mechanism, which mimics vulture behavior during the search process. The findings underscore the algorithm’s robustness, ease of use, and lack of dependence on derivative information. The review also critically evaluates the AVOA’s convergence behavior, identifying its strengths and limitations. In conclusion, the study not only consolidates the existing knowledge on AVOA but also proposes directions for future research, including potential adaptations and enhancements to address its limitations. The insights gained from this review offer valuable guidance for researchers and practitioners seeking to apply or improve the AVOA in various optimization tasks.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":null,"pages":null},"PeriodicalIF":10.7,"publicationDate":"2024-10-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10981-2.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142443224","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-17DOI: 10.1007/s10462-024-10928-7
Marco Gavanelli, Pascual Julián-Iranzo, Fernando Sáenz-Pérez
Abductive logic programming (ALP) extends logic programming with hypothetical reasoning by means of abducibles, an extension able to handle interesting problems, such as diagnosis, planning, and verification with formal methods. Implementations of this extension have been using Prolog meta-interpreters and Prolog programs with Constraint Handling Rules (CHR). While the latter adds a clean and efficient interface to the host system, it still suffers in performance for large programs. Here, the concern is to obtain a more performant implementation of the SCIFF system following a compiled approach. This paper, as a first step in this long term goal, sets out a propositional ALP system following SCIFF, eliminating the need for CHR and achieving better performance.
{"title":"An efficient propositional system for Abductive Logic Programming","authors":"Marco Gavanelli, Pascual Julián-Iranzo, Fernando Sáenz-Pérez","doi":"10.1007/s10462-024-10928-7","DOIUrl":"10.1007/s10462-024-10928-7","url":null,"abstract":"<div><p>Abductive logic programming (ALP) extends logic programming with hypothetical reasoning by means of abducibles, an extension able to handle interesting problems, such as diagnosis, planning, and verification with formal methods. Implementations of this extension have been using Prolog meta-interpreters and Prolog programs with Constraint Handling Rules (<span>CHR</span>). While the latter adds a clean and efficient interface to the host system, it still suffers in performance for large programs. Here, the concern is to obtain a more performant implementation of the <span>SCIFF</span> system following a compiled approach. This paper, as a first step in this long term goal, sets out a propositional ALP system following <span>SCIFF</span>, eliminating the need for <span>CHR</span> and achieving better performance.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":null,"pages":null},"PeriodicalIF":10.7,"publicationDate":"2024-10-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10928-7.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142443351","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-17DOI: 10.1007/s10462-024-10946-5
Gang Hu, Yuxuan Guo, Weiguo Zhao, Essam H. Houssein
In response to the shortcomings of particle swarm optimization (PSO), such as low execution efficiency and difficulty in overcoming local optima, this paper proposes a multi-strategy PSO method incorporating snow ablation operation (SAO), known as SAO-MPSO. Firstly, Cubic initialization is performed on particles to obtain a good initial environment. Subsequently, SAO and PSO are combined in parallel, and a balanced search mechanism led by multiple sub-populations is devised, significantly improving the search efficiency of overall population. Finally, the degree day method of SAO is introduced, and particles are endowed with memory of environmental changes to prevent premature convergence of PSO, while balancing the exploration and exploitation (ENE) capabilities in later phases. All adaptive parameters are used throughout this method in place of fixed parameters to improve the robustness and adaptability. For a comprehensive analysis of SAO-MPSO, its good ENE ability is verified on CEC 2020 and CEC 2022 and this method is compared with existing improved PSO versions on both test sets. The results show that SAO-MPSO has certain advantages in the comparison of similar improved algorithms. In order to further validate the strength of SAO-MPSO in dealing with nonlinear optimization problems (OPs) with strong constraints, firstly, based on the ball Wang-Ball (BWB) curve, a combined BWB (CBWB) curve is constructed, and a construction method for CBWB curves that satisfy G1 and G2 continuity is derived. Then, with the energy minimization and scale parameters of the CBWB curve as the optimization objective and variables respectively, a shape optimization model that satisfies G2 continuity is established. Finally, three numerical optimization examples based on this model are solved using SAO-MPSO and compared with 10 other methods. The results show that the energy obtained by SAO-MPSO is the smallest, which verifies the effectiveness of this method applied to shape OPs of CBWB curve.
{"title":"An adaptive snow ablation-inspired particle swarm optimization with its application in geometric optimization","authors":"Gang Hu, Yuxuan Guo, Weiguo Zhao, Essam H. Houssein","doi":"10.1007/s10462-024-10946-5","DOIUrl":"10.1007/s10462-024-10946-5","url":null,"abstract":"<div><p>In response to the shortcomings of particle swarm optimization (PSO), such as low execution efficiency and difficulty in overcoming local optima, this paper proposes a multi-strategy PSO method incorporating snow ablation operation (SAO), known as SAO-MPSO. Firstly, Cubic initialization is performed on particles to obtain a good initial environment. Subsequently, SAO and PSO are combined in parallel, and a balanced search mechanism led by multiple sub-populations is devised, significantly improving the search efficiency of overall population. Finally, the degree day method of SAO is introduced, and particles are endowed with memory of environmental changes to prevent premature convergence of PSO, while balancing the exploration and exploitation (ENE) capabilities in later phases. All adaptive parameters are used throughout this method in place of fixed parameters to improve the robustness and adaptability. For a comprehensive analysis of SAO-MPSO, its good ENE ability is verified on CEC 2020 and CEC 2022 and this method is compared with existing improved PSO versions on both test sets. The results show that SAO-MPSO has certain advantages in the comparison of similar improved algorithms. In order to further validate the strength of SAO-MPSO in dealing with nonlinear optimization problems (OPs) with strong constraints, firstly, based on the ball Wang-Ball (BWB) curve, a combined BWB (CBWB) curve is constructed, and a construction method for CBWB curves that satisfy <i>G</i><sup>1</sup> and <i>G</i><sup>2</sup> continuity is derived. Then, with the energy minimization and scale parameters of the CBWB curve as the optimization objective and variables respectively, a shape optimization model that satisfies <i>G</i><sup>2</sup> continuity is established. Finally, three numerical optimization examples based on this model are solved using SAO-MPSO and compared with 10 other methods. The results show that the energy obtained by SAO-MPSO is the smallest, which verifies the effectiveness of this method applied to shape OPs of CBWB curve.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":null,"pages":null},"PeriodicalIF":10.7,"publicationDate":"2024-10-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10946-5.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142443317","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-16DOI: 10.1007/s10462-024-10972-3
Marek Pawlicki, Aleksandra Pawlicka, Rafał Kozik, Michał Choraś
In the rapidly evolving domain of cybersecurity, the imperative for intrusion detection systems is undeniable; yet, it is increasingly clear that to meet the ever-growing challenges posed by sophisticated threats, intrusion detection itself stands in need of the transformative capabilities offered by the explainable artificial intelligence (xAI). As this concept is still developing, it poses an array of challenges that need addressing. This paper discusses 25 of such challenges of varying research interest, encountered in the domain of xAI, identified in the course of a targeted study. While these challenges may appear as obstacles, they concurrently present as significant research opportunities. These analysed challenges encompass a wide spectrum of concerns spanning the intersection of xAI and cybersecurity. The paper underscores the critical role of xAI in addressing opacity issues within machine learning algorithms and sets the stage for further research and innovation in the quest for transparent and interpretable artificial intelligence that humans are able to trust. In addition to this, by reframing these challenges as opportunities, this study seeks to inspire and guide researchers towards realizing the full potential of xAI in cybersecurity.
{"title":"The survey on the dual nature of xAI challenges in intrusion detection and their potential for AI innovation","authors":"Marek Pawlicki, Aleksandra Pawlicka, Rafał Kozik, Michał Choraś","doi":"10.1007/s10462-024-10972-3","DOIUrl":"10.1007/s10462-024-10972-3","url":null,"abstract":"<div><p>In the rapidly evolving domain of cybersecurity, the imperative for intrusion detection systems is undeniable; yet, it is increasingly clear that to meet the ever-growing challenges posed by sophisticated threats, intrusion detection itself stands in need of the transformative capabilities offered by the explainable artificial intelligence (xAI). As this concept is still developing, it poses an array of challenges that need addressing. This paper discusses 25 of such challenges of varying research interest, encountered in the domain of xAI, identified in the course of a targeted study. While these challenges may appear as obstacles, they concurrently present as significant research opportunities. These analysed challenges encompass a wide spectrum of concerns spanning the intersection of xAI and cybersecurity. The paper underscores the critical role of xAI in addressing opacity issues within machine learning algorithms and sets the stage for further research and innovation in the quest for transparent and interpretable artificial intelligence that humans are able to trust. In addition to this, by reframing these challenges as opportunities, this study seeks to inspire and guide researchers towards realizing the full potential of xAI in cybersecurity.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":null,"pages":null},"PeriodicalIF":10.7,"publicationDate":"2024-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10972-3.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142438878","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Alzheimer’s disease (AD) is a growing global concern, exacerbated by an aging population and the high costs associated with traditional detection methods. Recent research has identified speech data as valuable clinical information for AD detection, given its association with the progressive degeneration of brain cells and subsequent impacts on memory, cognition, and language abilities. The ongoing demographic shift toward an aging global population underscores the critical need for affordable and easily available methods for early AD detection and intervention. To address this major challenge, substantial research has recently focused on investigating speech data, aiming to develop efficient and affordable diagnostic tools that align with the demands of our aging society. This paper presents an in-depth review of studies from 2018–2023 utilizing speech for AD detection. Following the PRISMA protocol and a two-stage selection process, we identified 85 publications for analysis. In contrast to previous literature reviews, this paper places a strong emphasis on conducting a rigorous comparative analysis of various Artificial Intelligence (AI) based techniques, categorizing them meticulously based on underlying algorithms. We perform an exhaustive evaluation of research papers leveraging common benchmark datasets, specifically ADReSS and ADReSSo, to assess their performance. In contrast to previous literature reviews, this work makes a significant contribution by overcoming the limitations posed by the absence of standardized tasks and commonly accepted benchmark datasets for comparing different studies. The analysis reveals the dominance of deep learning models, particularly those leveraging pre-trained models like BERT, in AD detection. The integration of acoustic and linguistic features often achieves accuracies above 85%. Despite these advancements, challenges persist in data scarcity, standardization, privacy, and model interpretability. Future directions include improving multilingual recognition, exploring emerging multimodal approaches, and enhancing ASR systems for AD patients. By identifying these key challenges and suggesting future research directions, our review serves as a valuable resource for advancing AD detection techniques and their practical implementation.
{"title":"Speech based detection of Alzheimer’s disease: a survey of AI techniques, datasets and challenges","authors":"Kewen Ding, Madhu Chetty, Azadeh Noori Hoshyar, Tanusri Bhattacharya, Britt Klein","doi":"10.1007/s10462-024-10961-6","DOIUrl":"10.1007/s10462-024-10961-6","url":null,"abstract":"<div><p>Alzheimer’s disease (AD) is a growing global concern, exacerbated by an aging population and the high costs associated with traditional detection methods. Recent research has identified speech data as valuable clinical information for AD detection, given its association with the progressive degeneration of brain cells and subsequent impacts on memory, cognition, and language abilities. The ongoing demographic shift toward an aging global population underscores the critical need for affordable and easily available methods for early AD detection and intervention. To address this major challenge, substantial research has recently focused on investigating speech data, aiming to develop efficient and affordable diagnostic tools that align with the demands of our aging society. This paper presents an in-depth review of studies from 2018–2023 utilizing speech for AD detection. Following the PRISMA protocol and a two-stage selection process, we identified 85 publications for analysis. In contrast to previous literature reviews, this paper places a strong emphasis on conducting a rigorous comparative analysis of various Artificial Intelligence (AI) based techniques, categorizing them meticulously based on underlying algorithms. We perform an exhaustive evaluation of research papers leveraging common benchmark datasets, specifically ADReSS and ADReSSo, to assess their performance. In contrast to previous literature reviews, this work makes a significant contribution by overcoming the limitations posed by the absence of standardized tasks and commonly accepted benchmark datasets for comparing different studies. The analysis reveals the dominance of deep learning models, particularly those leveraging pre-trained models like BERT, in AD detection. The integration of acoustic and linguistic features often achieves accuracies above 85%. Despite these advancements, challenges persist in data scarcity, standardization, privacy, and model interpretability. Future directions include improving multilingual recognition, exploring emerging multimodal approaches, and enhancing ASR systems for AD patients. By identifying these key challenges and suggesting future research directions, our review serves as a valuable resource for advancing AD detection techniques and their practical implementation.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":null,"pages":null},"PeriodicalIF":10.7,"publicationDate":"2024-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10961-6.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142411527","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Effective task scheduling has become the key to optimizing resource allocation, reducing operation costs, and enhancing the user experience. The complexity and dynamics of cloud computing environments require task scheduling algorithms that can flexibly respond to multiple computing demands and changing resource states. Therefore, we propose an enhanced Red-tailed Hawk algorithm (named ERTH) based on multiple elite policies and chaotic mapping, while applying this approach in conjunction with the proposed scheduling model to optimize the efficiency of task scheduling in cloud computing environments. We apply the ERTH algorithm to a real cloud computing environment and conduct a comparison with the original RTH and other conventional algorithms. The proposed ERTH algorithm has better convergence speed and stability in most cases of small and large-scale tasks and performs better in minimizing the task completion time and system load cost. Specifically, our experiments show that the ERTH algorithm reduces the total system cost by 34.8% and 36.4% relative to the traditional algorithm for tasks of different sizes. Further, evaluations in the IEEE Congress on Evolutionary Computation (CEC) benchmark test sets show that the ERTH algorithm outperforms the traditional or emerging algorithms in several performance metrics such as mean, standard deviation, etc. The proposal and validation of the ERTH algorithm are of great significance in promoting the application of intelligent optimization algorithms in cloud computing.
{"title":"ERTH scheduler: enhanced red-tailed hawk algorithm for multi-cost optimization in cloud task scheduling","authors":"Xinqi Qin, Shaobo Li, Jian Tong, Cankun Xie, Xingxing Zhang, Fengbin Wu, Qun Xie, Yihong Ling, Guangzheng Lin","doi":"10.1007/s10462-024-10945-6","DOIUrl":"10.1007/s10462-024-10945-6","url":null,"abstract":"<div><p>Effective task scheduling has become the key to optimizing resource allocation, reducing operation costs, and enhancing the user experience. The complexity and dynamics of cloud computing environments require task scheduling algorithms that can flexibly respond to multiple computing demands and changing resource states. Therefore, we propose an enhanced Red-tailed Hawk algorithm (named ERTH) based on multiple elite policies and chaotic mapping, while applying this approach in conjunction with the proposed scheduling model to optimize the efficiency of task scheduling in cloud computing environments. We apply the ERTH algorithm to a real cloud computing environment and conduct a comparison with the original RTH and other conventional algorithms. The proposed ERTH algorithm has better convergence speed and stability in most cases of small and large-scale tasks and performs better in minimizing the task completion time and system load cost. Specifically, our experiments show that the ERTH algorithm reduces the total system cost by 34.8% and 36.4% relative to the traditional algorithm for tasks of different sizes. Further, evaluations in the IEEE Congress on Evolutionary Computation (CEC) benchmark test sets show that the ERTH algorithm outperforms the traditional or emerging algorithms in several performance metrics such as mean, standard deviation, etc. The proposal and validation of the ERTH algorithm are of great significance in promoting the application of intelligent optimization algorithms in cloud computing.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":null,"pages":null},"PeriodicalIF":10.7,"publicationDate":"2024-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10945-6.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142411499","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-12DOI: 10.1007/s10462-024-10858-4
Sandeep Kumar, Pranab K. Muhuri
Though economic growths of most of the nations have seen exponential rise due to industrialization, it has also caused proportional increase in their carbon emissions. This paper exploits this proportionate relationship of carbon emission with GDP to predict the per-capita GDP of those nations whose GDP values are missing in the world bank database. The reason behind the same was, those countries were either war-torn or politically isolated/unstable. To achieve the objective of predicting the missing GDP values of those countries from their carbon emissions, this paper exploits the non-linear relationship among the carbon emissions from solid fuels, liquid fuels, and gaseous fuels. It is so because even the differential utilization of these fuels impact economy differently. Use of traditional solid fuel for cooking points toward energy poverty, and access to clean cooking gas indicates higher living standard. However, the available data from the war-torn or isolated countries are very little, and hence insufficient for building a robust predictive machine learning model. So, this paper employs multi-source unsupervised transfer learning to precisely estimate the missing per-capita GDP of those nations. It suitably enlarges the training domains for the prediction models to be more robust. We empirically evaluate the proposed methodology for different regression techniques to estimate the missing GDP values of eleven different nations belonging to diverse strata of economies viz. developed economies, developing, and/or least developing economies. Proposed methodology profoundly improves the prediction preciseness of these regression techniques in estimating the missing per-capita GDP of the considered nations.
虽然大多数国家的经济增长因工业化而呈指数级增长,但这也导致了其碳排放量的成比例增加。本文利用碳排放量与国内生产总值的比例关系,预测那些在世界银行数据库中国内生产总值数值缺失的国家的人均国内生产总值。其背后的原因是,这些国家要么饱受战争蹂躏,要么政治孤立/不稳定。为了实现从这些国家的碳排放量预测其缺失的 GDP 值的目标,本文利用了固体燃料、液体燃料和气体燃料的碳排放量之间的非线性关系。这是因为即使这些燃料的利用率不同,对经济的影响也是不同的。使用传统固体燃料做饭表明能源贫困,而使用清洁燃气做饭则表明生活水平较高。然而,来自战乱国家或偏远国家的可用数据非常少,因此不足以建立一个强大的预测性机器学习模型。因此,本文采用多源无监督迁移学习来精确估算这些国家缺失的人均 GDP。它适当地扩大了预测模型的训练域,使其更加稳健。我们用不同的回归技术对所提出的方法进行了实证评估,以估算 11 个不同国家缺失的 GDP 值,这些国家属于不同的经济阶层,即发达经济体、发展中国家和/或最不发达经济体。在估算所考虑国家缺失的人均 GDP 时,所提出的方法大大提高了这些回归技术的预测精确度。
{"title":"A concise review towards a novel target specific multi-source unsupervised transfer learning technique for GDP estimation using CO2 emission data","authors":"Sandeep Kumar, Pranab K. Muhuri","doi":"10.1007/s10462-024-10858-4","DOIUrl":"10.1007/s10462-024-10858-4","url":null,"abstract":"<div><p>Though economic growths of most of the nations have seen exponential rise due to industrialization, it has also caused proportional increase in their carbon emissions. This paper exploits this proportionate relationship of carbon emission with GDP to predict the per-capita GDP of those nations whose GDP values are missing in the world bank database. The reason behind the same was, those countries were either war-torn or politically isolated/unstable. To achieve the objective of predicting the missing GDP values of those countries from their carbon emissions, this paper exploits the non-linear relationship among the carbon emissions from solid fuels, liquid fuels, and gaseous fuels. It is so because even the differential utilization of these fuels impact economy differently. Use of traditional solid fuel for cooking points toward energy poverty, and access to clean cooking gas indicates higher living standard. However, the available data from the war-torn or isolated countries are very little, and hence insufficient for building a robust predictive machine learning model. So, this paper employs multi-source unsupervised transfer learning to precisely estimate the missing per-capita GDP of those nations. It suitably enlarges the training domains for the prediction models to be more robust. We empirically evaluate the proposed methodology for different regression techniques to estimate the missing GDP values of eleven different nations belonging to diverse strata of economies viz. developed economies, developing, and/or least developing economies. Proposed methodology profoundly improves the prediction preciseness of these regression techniques in estimating the missing per-capita GDP of the considered nations.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":null,"pages":null},"PeriodicalIF":10.7,"publicationDate":"2024-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10858-4.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142411521","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This review paper focuses on the progress of deep learning-based methods for multi-view 3D object recognition. It covers the state-of-the-art techniques in this field, specifically those that utilize 3D multi-view data as input representation. The paper provides a comprehensive analysis of the pipeline for deep learning-based multi-view 3D object recognition, including the various techniques employed at each stage. It also presents the latest developments in CNN-based and transformer-based models for multi-view 3D object recognition. The review discusses existing models in detail, including the datasets, camera configurations, view selection strategies, pre-trained CNN architectures, fusion strategies, and recognition performance. Additionally, it examines various computer vision applications that use multi-view classification. Finally, it highlights future directions, factors impacting recognition performance, and trends for the development of multi-view 3D object recognition method.
本综述论文重点介绍基于深度学习的多视角三维物体识别方法的进展。它涵盖了该领域最先进的技术,特别是那些利用三维多视角数据作为输入表示的技术。论文全面分析了基于深度学习的多视角三维物体识别流程,包括每个阶段采用的各种技术。论文还介绍了基于 CNN 和变换器的多视角 3D 物体识别模型的最新发展。综述详细讨论了现有模型,包括数据集、相机配置、视图选择策略、预训练 CNN 架构、融合策略和识别性能。此外,它还研究了使用多视图分类的各种计算机视觉应用。最后,它强调了多视角三维物体识别方法的未来发展方向、影响识别性能的因素和发展趋势。
{"title":"Deep models for multi-view 3D object recognition: a review","authors":"Mona Alzahrani, Muhammad Usman, Salma Kammoun Jarraya, Saeed Anwar, Tarek Helmy","doi":"10.1007/s10462-024-10941-w","DOIUrl":"10.1007/s10462-024-10941-w","url":null,"abstract":"<div><p>This review paper focuses on the progress of deep learning-based methods for multi-view 3D object recognition. It covers the state-of-the-art techniques in this field, specifically those that utilize 3D multi-view data as input representation. The paper provides a comprehensive analysis of the pipeline for deep learning-based multi-view 3D object recognition, including the various techniques employed at each stage. It also presents the latest developments in CNN-based and transformer-based models for multi-view 3D object recognition. The review discusses existing models in detail, including the datasets, camera configurations, view selection strategies, pre-trained CNN architectures, fusion strategies, and recognition performance. Additionally, it examines various computer vision applications that use multi-view classification. Finally, it highlights future directions, factors impacting recognition performance, and trends for the development of multi-view 3D object recognition method.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":null,"pages":null},"PeriodicalIF":10.7,"publicationDate":"2024-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10941-w.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142411522","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-12DOI: 10.1007/s10462-024-10963-4
P. Alirezazadeh, F. Dornaika, J. Charafeddine
To enhance the accuracy of breast cancer diagnosis, current practices rely on biopsies and microscopic examinations. However, this approach is known for being time-consuming, tedious, and costly. While convolutional neural networks (CNNs) have shown promise for their efficiency and high accuracy, training them effectively becomes challenging in real-world learning scenarios such as class imbalance, small-scale datasets, and label noises. Angular margin-based softmax losses, which concentrate on the angle between features and classifiers embedded in cosine similarity at the classification layer, aim to regulate feature representation learning. Nevertheless, the cosine similarity’s lack of a heavy tail impedes its ability to compactly regulate intra-class feature distribution, limiting generalization performance. Moreover, these losses are constrained to target classes when margin penalties are applied, which may not always optimize effectiveness. Addressing these hurdles, we introduce an innovative approach termed MF-BAM (Mises-Fisher Similarity-based Boosted Additive Angular Margin Loss), which extends beyond traditional cosine similarity and is anchored in the von Mises-Fisher distribution. MF-BAM not only penalizes the angle between deep features and their corresponding target class weights but also considers angles between deep features and weights associated with non-target classes. Through extensive experimentation on the BreaKHis dataset, MF-BAM achieves outstanding accuracies of 99.92%, 99.96%, 100.00%, and 98.05% for magnification levels of ×40, ×100, ×200, and ×400, respectively. Furthermore, additional experiments conducted on the BACH dataset for breast cancer classification, as well as on the LFW and YTF datasets for face recognition, affirm the generalization capability of our proposed loss function.
{"title":"Mises-Fisher similarity-based boosted additive angular margin loss for breast cancer classification","authors":"P. Alirezazadeh, F. Dornaika, J. Charafeddine","doi":"10.1007/s10462-024-10963-4","DOIUrl":"10.1007/s10462-024-10963-4","url":null,"abstract":"<div><p>To enhance the accuracy of breast cancer diagnosis, current practices rely on biopsies and microscopic examinations. However, this approach is known for being time-consuming, tedious, and costly. While convolutional neural networks (CNNs) have shown promise for their efficiency and high accuracy, training them effectively becomes challenging in real-world learning scenarios such as class imbalance, small-scale datasets, and label noises. Angular margin-based softmax losses, which concentrate on the angle between features and classifiers embedded in cosine similarity at the classification layer, aim to regulate feature representation learning. Nevertheless, the cosine similarity’s lack of a heavy tail impedes its ability to compactly regulate intra-class feature distribution, limiting generalization performance. Moreover, these losses are constrained to target classes when margin penalties are applied, which may not always optimize effectiveness. Addressing these hurdles, we introduce an innovative approach termed MF-BAM (Mises-Fisher Similarity-based Boosted Additive Angular Margin Loss), which extends beyond traditional cosine similarity and is anchored in the von Mises-Fisher distribution. MF-BAM not only penalizes the angle between deep features and their corresponding target class weights but also considers angles between deep features and weights associated with non-target classes. Through extensive experimentation on the BreaKHis dataset, MF-BAM achieves outstanding accuracies of 99.92%, 99.96%, 100.00%, and 98.05% for magnification levels of ×40, ×100, ×200, and ×400, respectively. Furthermore, additional experiments conducted on the BACH dataset for breast cancer classification, as well as on the LFW and YTF datasets for face recognition, affirm the generalization capability of our proposed loss function.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":null,"pages":null},"PeriodicalIF":10.7,"publicationDate":"2024-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-024-10963-4.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142411474","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}