Pub Date : 2024-10-10DOI: 10.1109/TAI.2024.3477457
Xiaotian Song;Xiangning Xie;Zeqiong Lv;Gary G. Yen;Weiping Ding;Jiancheng Lv;Yanan Sun
Neural architecture search (NAS) has received increasing attention because of its exceptional merits in automating the design of deep neural network (DNN) architectures. However, the performance evaluation process, as a key part of NAS, often requires training a large number of DNNs. This inevitably makes NAS computationally expensive. In past years, many efficient evaluation methods (EEMs) have been proposed to address this critical issue. In this article, we comprehensively survey these EEMs published up to date, and provide a detailed analysis to motivate the further development of this research direction. Specifically, we divide the existing EEMs into four categories based on the number of DNNs trained for constructing these EEMs. The categorization can reflect the degree of efficiency in principle, which can in turn help quickly grasp the methodological features. In surveying each category, we further discuss the design principles and analyze the strengths and weaknesses to clarify the landscape of existing EEMs, thus making easily understanding the research trends of EEMs. Furthermore, we also discuss the current challenges and issues to identify future research directions in this emerging topic. In summary, this survey provides a convenient overview of EEM for interested users, and they can easily select the proper EEM method for the tasks at hand. In addition, the researchers in the NAS field could continue exploring the future directions suggested in the article.
{"title":"Efficient Evaluation Methods for Neural Architecture Search: A Survey","authors":"Xiaotian Song;Xiangning Xie;Zeqiong Lv;Gary G. Yen;Weiping Ding;Jiancheng Lv;Yanan Sun","doi":"10.1109/TAI.2024.3477457","DOIUrl":"https://doi.org/10.1109/TAI.2024.3477457","url":null,"abstract":"Neural architecture search (NAS) has received increasing attention because of its exceptional merits in automating the design of deep neural network (DNN) architectures. However, the performance evaluation process, as a key part of NAS, often requires training a large number of DNNs. This inevitably makes NAS computationally expensive. In past years, many efficient evaluation methods (EEMs) have been proposed to address this critical issue. In this article, we comprehensively survey these EEMs published up to date, and provide a detailed analysis to motivate the further development of this research direction. Specifically, we divide the existing EEMs into four categories based on the number of DNNs trained for constructing these EEMs. The categorization can reflect the degree of efficiency in principle, which can in turn help quickly grasp the methodological features. In surveying each category, we further discuss the design principles and analyze the strengths and weaknesses to clarify the landscape of existing EEMs, thus making easily understanding the research trends of EEMs. Furthermore, we also discuss the current challenges and issues to identify future research directions in this emerging topic. In summary, this survey provides a convenient overview of EEM for interested users, and they can easily select the proper EEM method for the tasks at hand. In addition, the researchers in the NAS field could continue exploring the future directions suggested in the article.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"5 12","pages":"5990-6011"},"PeriodicalIF":0.0,"publicationDate":"2024-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142810406","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-10DOI: 10.1109/TAI.2024.3477425
Arya Agrawal;Teena Sharma;Nishchal K. Verma
Real-time 3-D view reconstruction in an unfamiliar environment poses complexity for various applications due to varying conditions such as occlusion, latency, precision, etc. This article thoroughly examines and tests contemporary methodologies addressing challenges in 3-D view reconstruction. The methods being explored in this article are categorized into volumetric and mesh, generative adversarial network based, and open source library based methods. The exploration of these methods undergoes detailed discussions, encompassing methods, advantages, limitations, and empirical results. The real-time testing of each method is done on benchmarked datasets, including ShapeNet, Pascal 3D+, Pix3D, etc. The narrative highlights the crucial role of 3-D view reconstruction in domains such as robotics, virtual and augmented reality, medical imaging, cultural heritage preservation, etc. The article also anticipates future scopes by exploring generative models, unsupervised learning, and advanced sensor fusion to increase the robustness of the algorithms.
{"title":"A Comprehensive Exploration of Real-Time 3-D View Reconstruction Methods","authors":"Arya Agrawal;Teena Sharma;Nishchal K. Verma","doi":"10.1109/TAI.2024.3477425","DOIUrl":"https://doi.org/10.1109/TAI.2024.3477425","url":null,"abstract":"Real-time 3-D view reconstruction in an unfamiliar environment poses complexity for various applications due to varying conditions such as occlusion, latency, precision, etc. This article thoroughly examines and tests contemporary methodologies addressing challenges in 3-D view reconstruction. The methods being explored in this article are categorized into volumetric and mesh, generative adversarial network based, and open source library based methods. The exploration of these methods undergoes detailed discussions, encompassing methods, advantages, limitations, and empirical results. The real-time testing of each method is done on benchmarked datasets, including ShapeNet, Pascal 3D+, Pix3D, etc. The narrative highlights the crucial role of 3-D view reconstruction in domains such as robotics, virtual and augmented reality, medical imaging, cultural heritage preservation, etc. The article also anticipates future scopes by exploring generative models, unsupervised learning, and advanced sensor fusion to increase the robustness of the algorithms.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"5 12","pages":"5915-5927"},"PeriodicalIF":0.0,"publicationDate":"2024-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142810371","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-10DOI: 10.1109/TAI.2024.3478191
Kaixiang Yang;Wuxing Chen;Yifan Shi;Zhiwen Yu;C. L. Philip Chen
In the field of intelligent manufacturing, tackling the classification challenges caused by imbalanced data is crucial. Although the broad learning system (BLS) has been recognized as an effective and efficient method, its performance wanes with imbalanced datasets. Therefore, this article proposes a novel approach named simplified kernel-based cost-sensitive broad learning system (SKCSBLS) to address these issues. Based on the framework of cost-sensitive broad learning system (CSBLS) that assigns distinctive adjustment costs for individual classes, SKCSBLS emphasizes the importance of the minority class while mitigating the impact of data imbalance. Additionally, considering the complexity introduced by noisy or overlapping data points, we incorporate kernel mapping into the CSBLS. This improvement not only improves the system's capability to handle overlapping classes of samples, but also improves the overall classification effectiveness. Our experimental results highlight the potential of SKCSBLS in overcoming the challenges inherent in unbalanced data, providing a robust solution for advanced fault diagnosis in intelligent systems.
{"title":"Simplified Kernel-Based Cost-Sensitive Broad Learning System for Imbalanced Fault Diagnosis","authors":"Kaixiang Yang;Wuxing Chen;Yifan Shi;Zhiwen Yu;C. L. Philip Chen","doi":"10.1109/TAI.2024.3478191","DOIUrl":"https://doi.org/10.1109/TAI.2024.3478191","url":null,"abstract":"In the field of intelligent manufacturing, tackling the classification challenges caused by imbalanced data is crucial. Although the broad learning system (BLS) has been recognized as an effective and efficient method, its performance wanes with imbalanced datasets. Therefore, this article proposes a novel approach named simplified kernel-based cost-sensitive broad learning system (SKCSBLS) to address these issues. Based on the framework of cost-sensitive broad learning system (CSBLS) that assigns distinctive adjustment costs for individual classes, SKCSBLS emphasizes the importance of the minority class while mitigating the impact of data imbalance. Additionally, considering the complexity introduced by noisy or overlapping data points, we incorporate kernel mapping into the CSBLS. This improvement not only improves the system's capability to handle overlapping classes of samples, but also improves the overall classification effectiveness. Our experimental results highlight the potential of SKCSBLS in overcoming the challenges inherent in unbalanced data, providing a robust solution for advanced fault diagnosis in intelligent systems.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"5 12","pages":"6629-6644"},"PeriodicalIF":0.0,"publicationDate":"2024-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142825950","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-10DOI: 10.1109/TAI.2024.3477436
Xiaojing Zhang;Shuangrong Liu;Lin Wang;Bo Yang;Jiawei Fan
In this study, the adaptive hypersphere nearest neighbors (ASNN) method is proposed as an optimization framework to enhance the generalization performance of neural network classifiers. In terms of the classification task, the neural network draws decision boundaries by constructing the discriminative features of samples. To learn those features, attributed to the flexibility and separability, the pair-wise constraint-based methods that consist of the pair-wise loss and an embedding space (e.g., hypersphere space) have gained considerable attention over the last decade. Despite their success, pair-wise constraint-based methods still suffer from premature convergence or divergence problems, driven by two main challenges. 1) The poor scalability of the embedding space constrains the variety of the distribution of embedded samples, thereby increasing the optimization difficulty. 2) It is hard to select suitable positive/negative pairs during the training. In order to address the aforementioned problems, we propose an adaptive hypersphere nearest neighbors method. On the one hand, we improve the scalability of features via a scale-adaptive hypersphere embedding space. On the other hand, we introduce a neighborhood-based probability loss, which magnifies the difference between pairs and enhances the discriminative power of features generated by the neural networks based on the nearest neighbor-based pairing strategy. Experiments on UCI datasets and image recognition tasks demonstrate that the proposed ASNN not only achieves improved intraclass consistency and interclass separability of samples, but also outperforms its competitive counterparts.
{"title":"Learning Neural Network Classifiers by Distributing Nearest Neighbors on Adaptive Hypersphere","authors":"Xiaojing Zhang;Shuangrong Liu;Lin Wang;Bo Yang;Jiawei Fan","doi":"10.1109/TAI.2024.3477436","DOIUrl":"https://doi.org/10.1109/TAI.2024.3477436","url":null,"abstract":"In this study, the adaptive hypersphere nearest neighbors (ASNN) method is proposed as an optimization framework to enhance the generalization performance of neural network classifiers. In terms of the classification task, the neural network draws decision boundaries by constructing the discriminative features of samples. To learn those features, attributed to the flexibility and separability, the pair-wise constraint-based methods that consist of the pair-wise loss and an embedding space (e.g., hypersphere space) have gained considerable attention over the last decade. Despite their success, pair-wise constraint-based methods still suffer from premature convergence or divergence problems, driven by two main challenges. 1) The poor scalability of the embedding space constrains the variety of the distribution of embedded samples, thereby increasing the optimization difficulty. 2) It is hard to select suitable positive/negative pairs during the training. In order to address the aforementioned problems, we propose an adaptive hypersphere nearest neighbors method. On the one hand, we improve the scalability of features via a scale-adaptive hypersphere embedding space. On the other hand, we introduce a neighborhood-based probability loss, which magnifies the difference between pairs and enhances the discriminative power of features generated by the neural networks based on the nearest neighbor-based pairing strategy. Experiments on UCI datasets and image recognition tasks demonstrate that the proposed ASNN not only achieves improved intraclass consistency and interclass separability of samples, but also outperforms its competitive counterparts.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"6 1","pages":"234-249"},"PeriodicalIF":0.0,"publicationDate":"2024-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142976088","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-08DOI: 10.1109/TAI.2024.3476434
Jing Li;Jichen Wang;Zerui Li;Yu Kang;Wenjun Lv
Lithology identification plays a pivotal role in stratigraphic characterization and reservoir exploration. The promising field of intelligent logging lithology identification, which employs machine learning algorithms to infer lithology from logging curves, is gaining significant attention. However, models trained on labeled wells currently face challenges in accurately predicting the lithologies of new unlabeled wells due to significant discrepancies in data distribution among different wells caused by the complex sedimentary environment and variations in logging equipment. Additionally, there is no guarantee that newly drilled wells share the same lithology classes as previously explored ones. Therefore, our research aims to leverage source logging and lithology data along with target logging data to train a model capable of directly discerning the lithologies of target wells. The challenges are centered around the disparities in data distribution and the lack of prior knowledge regarding potential lithology classes in the target well. To tackle these concerns, we have made concerted efforts: 1) proposing a novel lithology identification framework, sample transferability weighting based partial domain adaptation (ST-PDA), to effectively address the practical scenario of encountering an unknown label space in target wells; 2) designing a sample transferability weighting module to assign higher weights to shared-class samples, thus effectively mitigating the negative transfer caused by unshared-class source samples; 3) developing a module, convolutional neural network with integrated channel attention mechanism (CG ${}^{2}$