Pub Date : 2024-01-22DOI: 10.1007/s00607-023-01246-x
Hui Du, Cuntao Ma, Depeng Lu, Jingrui Liu
Heterogeneous graph representation learning has strong expressiveness when dealing with large-scale relational graph data, and its purpose is to effectively represent the semantic information and heterogeneous structure information of nodes in the graph. Current methods typically use shallow models to embed semantic information on low-order neighbor nodes in the graph, which prevents the complete retention of higher-order semantic feature information. To address this issue, this paper proposes a heterogeneous graph network for higher-order semantic enhancement called HHSE. Specifically, our model uses the identity mapping mechanism of residual attention at the node feature level to enhance the information representation of nodes in the hidden layer, and then utilizes two aggregation strategies to improve the retention of high-order semantic information. The semantic feature level aims to learn the semantic information of nodes in various meta path subgraphs. Extensive experiments on node classification and node clustering on three real-existing datasets show that the proposed approach makes practical improvements compared to the state-of-the-art methods. Besides, our method is applicable to large-scale heterogeneous graph representation learning.
{"title":"HHSE: heterogeneous graph neural network via higher-order semantic enhancement","authors":"Hui Du, Cuntao Ma, Depeng Lu, Jingrui Liu","doi":"10.1007/s00607-023-01246-x","DOIUrl":"https://doi.org/10.1007/s00607-023-01246-x","url":null,"abstract":"<p>Heterogeneous graph representation learning has strong expressiveness when dealing with large-scale relational graph data, and its purpose is to effectively represent the semantic information and heterogeneous structure information of nodes in the graph. Current methods typically use shallow models to embed semantic information on low-order neighbor nodes in the graph, which prevents the complete retention of higher-order semantic feature information. To address this issue, this paper proposes a heterogeneous graph network for higher-order semantic enhancement called HHSE. Specifically, our model uses the identity mapping mechanism of residual attention at the node feature level to enhance the information representation of nodes in the hidden layer, and then utilizes two aggregation strategies to improve the retention of high-order semantic information. The semantic feature level aims to learn the semantic information of nodes in various meta path subgraphs. Extensive experiments on node classification and node clustering on three real-existing datasets show that the proposed approach makes practical improvements compared to the state-of-the-art methods. Besides, our method is applicable to large-scale heterogeneous graph representation learning.</p>","PeriodicalId":10718,"journal":{"name":"Computing","volume":"35 1 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139552321","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-19DOI: 10.1007/s00607-023-01244-z
Abstract
In this study, we propose a novel methodology called Particle Dynamics Method (PDM) for identifying and quantifying influential nodes in complex networks. Inspired by Newton’s three laws of motion and the universal gravitation law, PDM is based on a mathematical programming method that leverages node degrees and shortest path lengths. Unlike traditional centrality measures, PDM is easily adaptable to different network sizes and models, making it a versatile tool for network analysis. Our updated version of PDM also considers the direction of each force, resulting in more reliable results. To evaluate PDM’s performance, we tested it on a set of benchmark networks with distinct characteristics and models. Our results demonstrate that PDM outperforms other methodologies in the literature, as removing the identified influential nodes results in a significant decrease in network efficiency and robustness. The key feature of PDM is its flexibility in defining distance, which can be adapted to various network types. For instance, in a transportation network, distance can be defined by the flow between nodes, while in an academic publication system, the quartile of the journal could be used. Our research not only demonstrates the effectiveness of PDM but also highlights the influence of universities in the higher education and global university ranking networks, shedding light on the dynamics of these networks. Our interdisciplinary work has significant potential for collaborations between optimization, physics, and network science. This study opens up avenues for future research, including the extension of PDM to multilayer networks and the generalization of the metrics of monolayer networks for this purpose.
{"title":"Quantifying influential nodes in complex networks using optimization and particle dynamics: a comparative study","authors":"","doi":"10.1007/s00607-023-01244-z","DOIUrl":"https://doi.org/10.1007/s00607-023-01244-z","url":null,"abstract":"<h3>Abstract</h3> <p>In this study, we propose a novel methodology called Particle Dynamics Method (PDM) for identifying and quantifying influential nodes in complex networks. Inspired by Newton’s three laws of motion and the universal gravitation law, PDM is based on a mathematical programming method that leverages node degrees and shortest path lengths. Unlike traditional centrality measures, PDM is easily adaptable to different network sizes and models, making it a versatile tool for network analysis. Our updated version of PDM also considers the direction of each force, resulting in more reliable results. To evaluate PDM’s performance, we tested it on a set of benchmark networks with distinct characteristics and models. Our results demonstrate that PDM outperforms other methodologies in the literature, as removing the identified influential nodes results in a significant decrease in network efficiency and robustness. The key feature of PDM is its flexibility in defining distance, which can be adapted to various network types. For instance, in a transportation network, distance can be defined by the flow between nodes, while in an academic publication system, the quartile of the journal could be used. Our research not only demonstrates the effectiveness of PDM but also highlights the influence of universities in the higher education and global university ranking networks, shedding light on the dynamics of these networks. Our interdisciplinary work has significant potential for collaborations between optimization, physics, and network science. This study opens up avenues for future research, including the extension of PDM to multilayer networks and the generalization of the metrics of monolayer networks for this purpose.</p>","PeriodicalId":10718,"journal":{"name":"Computing","volume":"10 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139498155","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-18DOI: 10.1007/s00607-023-01245-y
Abstract
To ensure the reliability and operational efficiency of the grid system, this paper proposes an unmanned aerial vehicle (UAV)-assisted Power Internet of Things (PIoT), which obtains real-time grid data through PIoT devices to support the management optimization of the grid system. Compared with traditional UAV-assisted communication networks, this paper enables data collection and energy transmission services for PIoT devices through UAVs. Firstly, the flight-hover-communication protocol is used. When the UAVs approach the target devices, they stop flying and remain hovering to provide services. The UAV selects full duplex mode in the hovering state, i.e., within the coverage area of the UAV, it can collect data from the target device while providing charging for other devices. Secondly, the UAVs can provide services to the required devices in sequence. Considering the priorities of the devices, both the data queue state and the energy pair state of network devices are considered comprehensively. Therefore, the optimization problem is constructed as a multi-objective optimization problem. First, the multi-objective optimization problem is transformed into a Markov process. Then, a multi-objective dynamic resource allocation algorithm based on reinforcement learning is proposed for solving the multi-objective optimization problem. The simulation results show that the proposed resource allocation scheme can effectively achieve a reasonable allocation of UAV resources, joint multi-objective optimization, and improved system performance.
{"title":"UAV-assisted wireless charging and data processing of power IoT devices","authors":"","doi":"10.1007/s00607-023-01245-y","DOIUrl":"https://doi.org/10.1007/s00607-023-01245-y","url":null,"abstract":"<h3>Abstract</h3> <p>To ensure the reliability and operational efficiency of the grid system, this paper proposes an unmanned aerial vehicle (UAV)-assisted Power Internet of Things (PIoT), which obtains real-time grid data through PIoT devices to support the management optimization of the grid system. Compared with traditional UAV-assisted communication networks, this paper enables data collection and energy transmission services for PIoT devices through UAVs. Firstly, the flight-hover-communication protocol is used. When the UAVs approach the target devices, they stop flying and remain hovering to provide services. The UAV selects full duplex mode in the hovering state, i.e., within the coverage area of the UAV, it can collect data from the target device while providing charging for other devices. Secondly, the UAVs can provide services to the required devices in sequence. Considering the priorities of the devices, both the data queue state and the energy pair state of network devices are considered comprehensively. Therefore, the optimization problem is constructed as a multi-objective optimization problem. First, the multi-objective optimization problem is transformed into a Markov process. Then, a multi-objective dynamic resource allocation algorithm based on reinforcement learning is proposed for solving the multi-objective optimization problem. The simulation results show that the proposed resource allocation scheme can effectively achieve a reasonable allocation of UAV resources, joint multi-objective optimization, and improved system performance.</p>","PeriodicalId":10718,"journal":{"name":"Computing","volume":"7 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-01-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139498308","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-16DOI: 10.1007/s00607-023-01251-0
G. Rajeswari, R. Arthi, K. Murugan
Wireless Sensor Networks (WSNs) often operate in hostile environments and are subject to frequent failures. Failure of multiple sensor nodes causes the network to split into disjoint segments, which leads to network partitioning. Federating these disjoint segments is necessary to prevent detrimental effects on WSN applications. This paper investigates a recovery strategy using mobile relay nodes (MD-carrier) for restoring network connectivity. The proposed MD-carrier Tour Planning (MDTP) approach restores network connectivity of partitioned WSNs with reduced tour length and latency. For this reason, failure nodes are identified, and disjoint segments are formed with the k-means algorithm. Then, the Analytic Hierarchy Process (AHP) and the Technique for Order of Preference by Similarity to Ideal Solution (TOPSIS) are used for the election of an AGgregator Node (AGN) for each segment. Furthermore, an algorithm for identifying sojourn locations is proposed, which coordinates the maximum number of AGNs. Choosing the sojourn locations is a challenging task in WSN since the incorrect selection of the sojourn locations would degrade its data collection process. This paper uses the nature-inspired meta-heuristic Donkey And Smuggler Optimization (DASO) algorithm to compute the optimal touring path. MDTP reduces tour length and latency by an average of 30.28% & 24.56% compared to existing approaches.
{"title":"Nature-inspired donkey and smuggler algorithm for optimal data gathering in partitioned wireless sensor networks for restoring network connectivity","authors":"G. Rajeswari, R. Arthi, K. Murugan","doi":"10.1007/s00607-023-01251-0","DOIUrl":"https://doi.org/10.1007/s00607-023-01251-0","url":null,"abstract":"<p>Wireless Sensor Networks (WSNs) often operate in hostile environments and are subject to frequent failures. Failure of multiple sensor nodes causes the network to split into disjoint segments, which leads to network partitioning. Federating these disjoint segments is necessary to prevent detrimental effects on WSN applications. This paper investigates a recovery strategy using mobile relay nodes (MD-carrier) for restoring network connectivity. The proposed MD-carrier Tour Planning (MDTP) approach restores network connectivity of partitioned WSNs with reduced tour length and latency. For this reason, failure nodes are identified, and disjoint segments are formed with the k-means algorithm. Then, the Analytic Hierarchy Process (AHP) and the Technique for Order of Preference by Similarity to Ideal Solution (TOPSIS) are used for the election of an AGgregator Node (AGN) for each segment. Furthermore, an algorithm for identifying sojourn locations is proposed, which coordinates the maximum number of AGNs. Choosing the sojourn locations is a challenging task in WSN since the incorrect selection of the sojourn locations would degrade its data collection process. This paper uses the nature-inspired meta-heuristic Donkey And Smuggler Optimization (DASO) algorithm to compute the optimal touring path. MDTP reduces tour length and latency by an average of 30.28% & 24.56% compared to existing approaches.</p>","PeriodicalId":10718,"journal":{"name":"Computing","volume":"3 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139482868","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-23DOI: 10.1007/s00607-023-01243-0
Armin Shoughi, Mohammad Bagher Dowlatshahi, Arefeh Amiri, Marjan Kuchaki Rafsanjani, Ranbir Singh Batth
This paper presents an approach based on deep learning for accurate Electrocardiogram signal classification. The electrocardiogram is a significant signal in the realm of medical affairs, which gives vital information about the cardiovascular status of patients to heart specialists. Manually meticulous analysis of signals needs high and specific skills, and it is a time-consuming job too. The existence of noise, the inflexibility of signals, and the irregularity of heartbeats keep heart specialists in trouble. Cardiovascular diseases (CVDs) are the most important factor of fatality globally, which annually caused the deaths of 17.9 million people. Totally 31% of all death in the world are related to CVDs, which the age of 1/3 of patients that died because of CVDs is below 70 Because of the high percentage of mortality in cardiovascular patients, accurate diagnosis of this disease is an important matter. We present an approach to the analysis of electrocardiogram signals based on the convolutional neural network, discrete wavelet transformation with db2 mother wavelet, and synthetic minority over-sampling technique (SMOTE) on the MIT-BIH dataset according to the association for the advancement of medical instrumentation (AAMI) standards to increase the accuracy in electrocardiogram signal classifications. The evaluation results show this approach with 50 epoch training that the time of each epoch is 39 s, achieved 99.71% accuracy for category F, 98.69% accuracy for category N, 99.45% accuracy for category S, 99.33% accuracy for category V and 99.82% accuracy for category Q. It is worth mentioning that it can potentially be used as a clinical auxiliary diagnostic tool. The source code is available at https://gitlab.com/arminshoughi/ecg-classification-cnn.
{"title":"Automatic ECG classification using discrete wavelet transform and one-dimensional convolutional neural network","authors":"Armin Shoughi, Mohammad Bagher Dowlatshahi, Arefeh Amiri, Marjan Kuchaki Rafsanjani, Ranbir Singh Batth","doi":"10.1007/s00607-023-01243-0","DOIUrl":"https://doi.org/10.1007/s00607-023-01243-0","url":null,"abstract":"<p>This paper presents an approach based on deep learning for accurate Electrocardiogram signal classification. The electrocardiogram is a significant signal in the realm of medical affairs, which gives vital information about the cardiovascular status of patients to heart specialists. Manually meticulous analysis of signals needs high and specific skills, and it is a time-consuming job too. The existence of noise, the inflexibility of signals, and the irregularity of heartbeats keep heart specialists in trouble. Cardiovascular diseases (CVDs) are the most important factor of fatality globally, which annually caused the deaths of 17.9 million people. Totally 31% of all death in the world are related to CVDs, which the age of 1/3 of patients that died because of CVDs is below 70 Because of the high percentage of mortality in cardiovascular patients, accurate diagnosis of this disease is an important matter. We present an approach to the analysis of electrocardiogram signals based on the convolutional neural network, discrete wavelet transformation with db2 mother wavelet, and synthetic minority over-sampling technique (SMOTE) on the MIT-BIH dataset according to the association for the advancement of medical instrumentation (AAMI) standards to increase the accuracy in electrocardiogram signal classifications. The evaluation results show this approach with 50 epoch training that the time of each epoch is 39 s, achieved 99.71% accuracy for category F, 98.69% accuracy for category N, 99.45% accuracy for category S, 99.33% accuracy for category V and 99.82% accuracy for category Q. It is worth mentioning that it can potentially be used as a clinical auxiliary diagnostic tool. The source code is available at https://gitlab.com/arminshoughi/ecg-classification-cnn.</p>","PeriodicalId":10718,"journal":{"name":"Computing","volume":"573 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2023-12-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139025398","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-21DOI: 10.1007/s00607-023-01222-5
B. Benatallah, Hakim Hacid, Eleana Kafeza, Fabio Martinelli, A. Bouguettaya
{"title":"ICSOC 2020 special issue on service-oriented computing","authors":"B. Benatallah, Hakim Hacid, Eleana Kafeza, Fabio Martinelli, A. Bouguettaya","doi":"10.1007/s00607-023-01222-5","DOIUrl":"https://doi.org/10.1007/s00607-023-01222-5","url":null,"abstract":"","PeriodicalId":10718,"journal":{"name":"Computing","volume":"79 9","pages":""},"PeriodicalIF":3.7,"publicationDate":"2023-12-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138951362","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-20DOI: 10.1007/s00607-023-01242-1
Mingxue Liao, Pin Lv
Some computations are based on structures of random partition. They take an n-size problem as input, then break this problem into sub-problems of randomized size, execute calculations on each sub-problems and combine results from these calculations at last. We propose a combinatorial method for analyzing such computations and prove that the averaged time complexity is in terms of Stirling numbers of the second kind. The result shows that the average time complexity is decreased about one order of magnitude compared to that of the original solution. We also show two application cases where random partition structures are applied to improve performance.
有些计算基于随机分区结构。它们将一个 n 大小的问题作为输入,然后将这个问题分解成随机大小的子问题,对每个子问题执行计算,最后将这些计算结果合并。我们提出了一种分析此类计算的组合方法,并证明了平均时间复杂度是以第二类斯特林数表示的。结果表明,与原始解法相比,平均时间复杂度降低了约一个数量级。我们还展示了应用随机分区结构提高性能的两个应用案例。
{"title":"On the average time complexity of computation with random partition","authors":"Mingxue Liao, Pin Lv","doi":"10.1007/s00607-023-01242-1","DOIUrl":"https://doi.org/10.1007/s00607-023-01242-1","url":null,"abstract":"<p>Some computations are based on structures of random partition. They take an <i>n</i>-size problem as input, then break this problem into sub-problems of randomized size, execute calculations on each sub-problems and combine results from these calculations at last. We propose a combinatorial method for analyzing such computations and prove that the averaged time complexity is in terms of Stirling numbers of the second kind. The result shows that the average time complexity is decreased about one order of magnitude compared to that of the original solution. We also show two application cases where random partition structures are applied to improve performance.</p>","PeriodicalId":10718,"journal":{"name":"Computing","volume":"245 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2023-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138819673","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-12DOI: 10.1007/s00607-023-01241-2
Weidong Li, Chunbo Shi, Yongbo Yu, Zhe Wang
The topological relationship of spatial data is essential to GIS data processing and spatial analysis such as in analysis of pipe explosion in gas pipeline network. The existing browser-side JavaScript topology check library is inefficient and even crashes when checking the pipe network topology relationships for large amounts of data. In this paper, we present a topology checking and optimization method for pipeline vector data in browser-side using quadtree. Firstly, an algorithm mechanism that conforms to GIS data is designed based on JavaScript shared memory mechanism, topological check algorithm characteristics, and spatial data high-precision characteristics. Then using a fast rejection experiment and straddle test to realize the browser-side topology checking algorithm, through tolerance setting, improve the inspection efficiency and accuracy, which solves the problem that Turf and Jsts libraries cannot set tolerance. Based on the concept of quadtree spatial index, an optimization method of browser-side quadtree topology checking algorithm(BQTCA) is proposed. Without setting tolerance, the topology check of 114 point data and 1881 line data takes 487 milliseconds, and the efficiency of BQTCA is about 12 times and 39 times higher than that of the well-known public libraries Turf and Jsts, respectively. When the data volume increases to 912 point data and 15048 line data, BQTCA takes 6970 ms, which is about 65 times and 190 times more efficient than Turf and Jsts, respectively. The larger the data volume is, the more pronounced the efficiency improvement of BQTCA. Even when the data volume is so large that Turf and Jsts can- not calculate even crash, BQTCA can still complete the checking calculation. Through experiments, BQTCA can significantly improve the efficiency of browser-side vector pipeline topology relationship inspection under a large amount of data, and meet the commercial application requirements.
{"title":"A novel optimization approach to topology checking of pipeline vector data in browser side","authors":"Weidong Li, Chunbo Shi, Yongbo Yu, Zhe Wang","doi":"10.1007/s00607-023-01241-2","DOIUrl":"https://doi.org/10.1007/s00607-023-01241-2","url":null,"abstract":"<p>The topological relationship of spatial data is essential to GIS data processing and spatial analysis such as in analysis of pipe explosion in gas pipeline network. The existing browser-side JavaScript topology check library is inefficient and even crashes when checking the pipe network topology relationships for large amounts of data. In this paper, we present a topology checking and optimization method for pipeline vector data in browser-side using quadtree. Firstly, an algorithm mechanism that conforms to GIS data is designed based on JavaScript shared memory mechanism, topological check algorithm characteristics, and spatial data high-precision characteristics. Then using a fast rejection experiment and straddle test to realize the browser-side topology checking algorithm, through tolerance setting, improve the inspection efficiency and accuracy, which solves the problem that Turf and Jsts libraries cannot set tolerance. Based on the concept of quadtree spatial index, an optimization method of browser-side quadtree topology checking algorithm(BQTCA) is proposed. Without setting tolerance, the topology check of 114 point data and 1881 line data takes 487 milliseconds, and the efficiency of BQTCA is about 12 times and 39 times higher than that of the well-known public libraries Turf and Jsts, respectively. When the data volume increases to 912 point data and 15048 line data, BQTCA takes 6970 ms, which is about 65 times and 190 times more efficient than Turf and Jsts, respectively. The larger the data volume is, the more pronounced the efficiency improvement of BQTCA. Even when the data volume is so large that Turf and Jsts can- not calculate even crash, BQTCA can still complete the checking calculation. Through experiments, BQTCA can significantly improve the efficiency of browser-side vector pipeline topology relationship inspection under a large amount of data, and meet the commercial application requirements.</p>","PeriodicalId":10718,"journal":{"name":"Computing","volume":"82 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2023-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138631068","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-11DOI: 10.1007/s00607-023-01240-3
Chuandong Qin, Yu Cao
At present, large-scale sky surveys have obtained a large volume of stellar spectra. An efficient classification algorithm is of great importance to the practice of astronomical research. In this paper, we propose a novel parallel optimization algorithm based on a lock-free and shared-memory environment to solve the model for astronomical spectra class. Firstly, the SMOTE-TOMEK and RobustScaler are introduced to use for class balancing and data normalization. Secondly, 1-Dimensional Convolutional Neural Networks (1-D CNN) with L2-norm loss function is utilized as a classifier. Finally, LFA-SGD, LFA-Adagrad, LFA-RMSprop and LFA-Adam algorithms are proposed and applied to the classifier solution. The Lock-Free and shared-memory parallel Asynchronous environment (LFA) relies on GPU multiprocessing, allowing the algorithm to fully utilize the multi-core resources of the computer. Due to its sparsity, the convergence speed is significantly faster. The experimental results show that LFA-SGD algorithm and its variants achieved state-of-the-art accuracy and efficiency for astronomical spectra class.
{"title":"1-D CNNs with lock-free asynchronous adaptive stochastic gradient descent algorithm for classification of astronomical spectra","authors":"Chuandong Qin, Yu Cao","doi":"10.1007/s00607-023-01240-3","DOIUrl":"https://doi.org/10.1007/s00607-023-01240-3","url":null,"abstract":"<p>At present, large-scale sky surveys have obtained a large volume of stellar spectra. An efficient classification algorithm is of great importance to the practice of astronomical research. In this paper, we propose a novel parallel optimization algorithm based on a lock-free and shared-memory environment to solve the model for astronomical spectra class. Firstly, the SMOTE-TOMEK and RobustScaler are introduced to use for class balancing and data normalization. Secondly, 1-Dimensional Convolutional Neural Networks (1-D CNN) with L2-norm loss function is utilized as a classifier. Finally, LFA-SGD, LFA-Adagrad, LFA-RMSprop and LFA-Adam algorithms are proposed and applied to the classifier solution. The Lock-Free and shared-memory parallel Asynchronous environment (LFA) relies on GPU multiprocessing, allowing the algorithm to fully utilize the multi-core resources of the computer. Due to its sparsity, the convergence speed is significantly faster. The experimental results show that LFA-SGD algorithm and its variants achieved state-of-the-art accuracy and efficiency for astronomical spectra class.</p>","PeriodicalId":10718,"journal":{"name":"Computing","volume":"59 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2023-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138566577","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}