The integration of different learning paradigms has long been a focus of machine learning research, aimed at overcoming the inherent limitations of individual methods. Fuzzy rule-based models excel in interpretability and have seen widespread application across diverse fields. However, they face challenges such as complex design specifications and scalability issues with large datasets. The fusion of different techniques and strategies, particularly gradient boosting, with fuzzy rule-based models offers a robust solution to these challenges. This article proposes an integrated fusion framework that merges the strengths of both paradigms to enhance model performance and interpretability. At each iteration, a fuzzy rule-based model is constructed and controlled by a dynamic factor to optimize its contribution to the overall ensemble. This control factor serves multiple purposes: it prevents model dominance, encourages diversity, acts as a regularization parameter, and provides a mechanism for dynamic tuning based on model performance, thus mitigating the risk of overfitting. Additionally, the framework incorporates a sample-based correction mechanism that allows for adaptive adjustments based on feedback from a validation set. Experimental results substantiate the efficacy of the presented gradient-boosting framework for fuzzy rule-based models, demonstrating performance enhancement, especially in terms of mitigating overfitting and complexity typically associated with many rules. By leveraging an optimal factor to govern the contribution of each model, the framework improves performance, maintains interpretability, and simplifies the maintenance and update of the models.
{"title":"An Integrated Fusion Framework for Ensemble Learning Leveraging Gradient-Boosting and Fuzzy Rule-Based Models","authors":"Jinbo Li;Peng Liu;Long Chen;Witold Pedrycz;Weiping Ding","doi":"10.1109/TAI.2024.3424427","DOIUrl":"https://doi.org/10.1109/TAI.2024.3424427","url":null,"abstract":"The integration of different learning paradigms has long been a focus of machine learning research, aimed at overcoming the inherent limitations of individual methods. Fuzzy rule-based models excel in interpretability and have seen widespread application across diverse fields. However, they face challenges such as complex design specifications and scalability issues with large datasets. The fusion of different techniques and strategies, particularly gradient boosting, with fuzzy rule-based models offers a robust solution to these challenges. This article proposes an integrated fusion framework that merges the strengths of both paradigms to enhance model performance and interpretability. At each iteration, a fuzzy rule-based model is constructed and controlled by a dynamic factor to optimize its contribution to the overall ensemble. This control factor serves multiple purposes: it prevents model dominance, encourages diversity, acts as a regularization parameter, and provides a mechanism for dynamic tuning based on model performance, thus mitigating the risk of overfitting. Additionally, the framework incorporates a sample-based correction mechanism that allows for adaptive adjustments based on feedback from a validation set. Experimental results substantiate the efficacy of the presented gradient-boosting framework for fuzzy rule-based models, demonstrating performance enhancement, especially in terms of mitigating overfitting and complexity typically associated with many rules. By leveraging an optimal factor to govern the contribution of each model, the framework improves performance, maintains interpretability, and simplifies the maintenance and update of the models.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"5 11","pages":"5771-5785"},"PeriodicalIF":0.0,"publicationDate":"2024-07-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142600170","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-05DOI: 10.1109/TAI.2024.3423813
Priyanti Paul Tumpa;Md. Saiful Islam
Satellite image classification is crucial for various applications, driving advancements in convolutional neural networks (CNNs). While CNNs have proven effective, deep models often encounter overfitting issues as the network's depth increases since the model has to learn many parameters. Besides this, traditional CNNs have the inherent difficulty of extracting fine-grained details and broader patterns simultaneously. To overcome these challenges, this article presents a novel approach using a lightweight parallel CNN (LPCNN) architecture with a support vector machine (SVM) classifier to classify satellite images. At first, preprocessing such as resizing and sharpening is used to improve image quality. Each branch within the parallel network is designed for specific resolution characteristics, spanning from low (emphasizing broader patterns) to high (capturing fine-grained details), enabling the simultaneous extraction of a comprehensive set of features without increasing network depth. The LPCNN incorporates a dilation factor to expand the network's receptive field without increasing parameters, and a dropout layer is introduced to mitigate overfitting. SVM is used alongside LPCNN because it is effective at handling high-dimensional features and defining complex decision boundaries, which improves overall classification accuracy. Evaluation of two public datasets (EuroSAT dataset and RSI-CB256 dataset) demonstrates remarkable accuracy rates of 97.91% and 99.8%, surpassing previous state-of-the-art models. Finally, LPCNN, with less than 1 million parameters, outperforms high-parameter models by effectively addressing overfitting issues, showcasing exceptional performance in satellite image classification.
{"title":"Lightweight Parallel Convolutional Neural Network With SVM Classifier for Satellite Imagery Classification","authors":"Priyanti Paul Tumpa;Md. Saiful Islam","doi":"10.1109/TAI.2024.3423813","DOIUrl":"https://doi.org/10.1109/TAI.2024.3423813","url":null,"abstract":"Satellite image classification is crucial for various applications, driving advancements in convolutional neural networks (CNNs). While CNNs have proven effective, deep models often encounter overfitting issues as the network's depth increases since the model has to learn many parameters. Besides this, traditional CNNs have the inherent difficulty of extracting fine-grained details and broader patterns simultaneously. To overcome these challenges, this article presents a novel approach using a lightweight parallel CNN (LPCNN) architecture with a support vector machine (SVM) classifier to classify satellite images. At first, preprocessing such as resizing and sharpening is used to improve image quality. Each branch within the parallel network is designed for specific resolution characteristics, spanning from low (emphasizing broader patterns) to high (capturing fine-grained details), enabling the simultaneous extraction of a comprehensive set of features without increasing network depth. The LPCNN incorporates a dilation factor to expand the network's receptive field without increasing parameters, and a dropout layer is introduced to mitigate overfitting. SVM is used alongside LPCNN because it is effective at handling high-dimensional features and defining complex decision boundaries, which improves overall classification accuracy. Evaluation of two public datasets (EuroSAT dataset and RSI-CB256 dataset) demonstrates remarkable accuracy rates of 97.91% and 99.8%, surpassing previous state-of-the-art models. Finally, LPCNN, with less than 1 million parameters, outperforms high-parameter models by effectively addressing overfitting issues, showcasing exceptional performance in satellite image classification.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"5 11","pages":"5676-5688"},"PeriodicalIF":0.0,"publicationDate":"2024-07-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142600283","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-03DOI: 10.1109/TAI.2024.3419749
Haoran Duan;Beibei Yu;Cheng Xie
Graph-structured data plays a foundational role in knowledge representation across various intelligent systems. Self-supervised graph representation learning (SSGRL) has emerged as a key methodology for processing such data efficiently. Recent advances in SSGRL have introduced the masked graph model (MGM), which achieves state-of-the-art performance by masking and reconstructing node features. However, the effectiveness of MGM-based methods heavily relies on the information density of the original node features. Performance deteriorates notably when dealing with sparse node features, such as one-hot and degree-hot encodings, commonly found in social and chemical graphs. To address this challenge, we propose a novel cross-view node feature reconstruction method that circumvents direct reliance on the original node features. Our approach generates four distinct views (graph view, masked view, diffusion view, and masked diffusion view) from the original graph through node masking and diffusion. These views are then encoded into representations with high information density. The reconstruction process operates across these representations, enabling self-supervised learning without direct reliance on the original features. Extensive experiments are conducted on 26 real-world graph datasets, including those with sparse and high information density environments. This cross-view reconstruction method represents a promising direction for effective SSGRL, particularly in scenarios with sparse node feature information.
{"title":"Cross-View Masked Model for Self-Supervised Graph Representation Learning","authors":"Haoran Duan;Beibei Yu;Cheng Xie","doi":"10.1109/TAI.2024.3419749","DOIUrl":"https://doi.org/10.1109/TAI.2024.3419749","url":null,"abstract":"Graph-structured data plays a foundational role in knowledge representation across various intelligent systems. Self-supervised graph representation learning (SSGRL) has emerged as a key methodology for processing such data efficiently. Recent advances in SSGRL have introduced the masked graph model (MGM), which achieves state-of-the-art performance by masking and reconstructing node features. However, the effectiveness of MGM-based methods heavily relies on the information density of the original node features. Performance deteriorates notably when dealing with sparse node features, such as one-hot and degree-hot encodings, commonly found in social and chemical graphs. To address this challenge, we propose a novel cross-view node feature reconstruction method that circumvents direct reliance on the original node features. Our approach generates four distinct views (graph view, masked view, diffusion view, and masked diffusion view) from the original graph through node masking and diffusion. These views are then encoded into representations with high information density. The reconstruction process operates across these representations, enabling self-supervised learning without direct reliance on the original features. Extensive experiments are conducted on 26 real-world graph datasets, including those with sparse and high information density environments. This cross-view reconstruction method represents a promising direction for effective SSGRL, particularly in scenarios with sparse node feature information.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"5 11","pages":"5540-5552"},"PeriodicalIF":0.0,"publicationDate":"2024-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142600167","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-02DOI: 10.1109/TAI.2024.3421175
Zihan Jiang;Yiqun Ma;Bingyu Shi;Xin Lu;Jian Xing;Nuno Gonçalves;Bo Jin
This article introduces a novel model for low-quality pedestrian trajectory prediction, the social nonstationary transformers (NSTransformers), that merges the strengths of NSTransformers and spatiotemporal graph transformer (STAR). The model can capture social interaction cues among pedestrians and integrate features across spatial and temporal dimensions to enhance the precision and resilience of trajectory predictions. We also propose an enhanced loss function that combines diversity loss with logarithmic root mean squared error (log-RMSE) to guarantee the reasonableness and diversity of the generated trajectories. This design adapts well to complex pedestrian interaction scenarios, thereby improving the reliability and accuracy of trajectory prediction. Furthermore, we integrate a generative adversarial network (GAN) to model the randomness inherent in pedestrian trajectories. Compared to the conventional standard Gaussian distribution, our GAN approach better simulates the intricate distribution found in pedestrian trajectories, enhancing the trajectory prediction's diversity and robustness. Experimental results reveal that our model outperforms several state-of-the-art methods. This research opens the avenue for future exploration in low-quality pedestrian trajectory prediction.
本文介绍了一种用于低质量行人轨迹预测的新型模型--社会非稳态变换器(NSTransformers),该模型融合了社会非稳态变换器和时空图变换器(STAR)的优点。该模型可以捕捉行人之间的社会互动线索,并整合跨时空维度的特征,从而提高轨迹预测的精度和弹性。我们还提出了一种增强型损失函数,将多样性损失与对数均方根误差(log-RMSE)相结合,以保证生成轨迹的合理性和多样性。这种设计能很好地适应复杂的行人交互场景,从而提高轨迹预测的可靠性和准确性。此外,我们还整合了生成式对抗网络(GAN)来模拟行人轨迹固有的随机性。与传统的标准高斯分布相比,我们的 GAN 方法能更好地模拟行人轨迹中错综复杂的分布,从而增强轨迹预测的多样性和鲁棒性。实验结果表明,我们的模型优于几种最先进的方法。这项研究为未来探索低质量行人轨迹预测开辟了道路。
{"title":"Social NSTransformers: Low-Quality Pedestrian Trajectory Prediction","authors":"Zihan Jiang;Yiqun Ma;Bingyu Shi;Xin Lu;Jian Xing;Nuno Gonçalves;Bo Jin","doi":"10.1109/TAI.2024.3421175","DOIUrl":"https://doi.org/10.1109/TAI.2024.3421175","url":null,"abstract":"This article introduces a novel model for low-quality pedestrian trajectory prediction, the social nonstationary transformers (NSTransformers), that merges the strengths of NSTransformers and spatiotemporal graph transformer (STAR). The model can capture social interaction cues among pedestrians and integrate features across spatial and temporal dimensions to enhance the precision and resilience of trajectory predictions. We also propose an enhanced loss function that combines diversity loss with logarithmic root mean squared error (log-RMSE) to guarantee the reasonableness and diversity of the generated trajectories. This design adapts well to complex pedestrian interaction scenarios, thereby improving the reliability and accuracy of trajectory prediction. Furthermore, we integrate a generative adversarial network (GAN) to model the randomness inherent in pedestrian trajectories. Compared to the conventional standard Gaussian distribution, our GAN approach better simulates the intricate distribution found in pedestrian trajectories, enhancing the trajectory prediction's diversity and robustness. Experimental results reveal that our model outperforms several state-of-the-art methods. This research opens the avenue for future exploration in low-quality pedestrian trajectory prediction.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"5 11","pages":"5575-5588"},"PeriodicalIF":0.0,"publicationDate":"2024-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142600184","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-02DOI: 10.1109/TAI.2024.3421172
Zhuoqing Liu;Tong Yang;Yongchun Fang;Ning Sun
With flexible payload adjustment ability and large load capacity, dual rotary cranes (DRCs) provide effective solutions for various complex hoisting tasks. At present, the control research for DRCs mostly focuses on two-dimensional space (restricting workspace and efficiency), or lacks the consideration of DRC dynamic characteristics and the practical demands for the dynamic regulation of payload positions and attitudes, which makes it difficult to handle hoisting tasks in complex environments. To tackle these issues, this article proposes an optimal trajectory-based motion control method for three-dimensional (3-D) DRCs in complex environments, effectively tackling key challenges encountered by DRCs operating in 3-D space. The proposed method achieves dynamic regulation of payload position and attitude by DRCs in 3-D space for the first