Pub Date : 2022-11-26DOI: 10.1109/ISCMI56532.2022.10068468
Raha Soleymanzadeh, R. Kashef
Various research studies have been recently introduced in developing generative models, especially in computer vision and image classification. These models are inspired by a generator and discriminator network architecture in a min-max optimization game called Generative Adversarial Networks (GANs). However, GANs-based models suffer from training instability, which means high oscillations during the training, which provides inaccurate results. There are various causes beyond the instability behaviours, such as the adopted generator architecture, loss function, and distance metrics. In this paper, we focus on the impact of the generator architectures and the loss functions on the GANs training. We aim to provide a comparative assessment of various architectures focusing on ensemble and hybrid models and loss functions such as Focal loss, Binary Cross-Entropy and Mean Squared loss function. Experimental results on NSL-KDD and UNSW-NB15 datasets show that the ensemble models are more stable in terms of training and have higher intrusion detection rates. Additionally, the focal loss can improve the performance of detection minority classes. Using Mean squared loss improved the detection rate for discriminator, however with the Binary Cross entropy loss function, the deep features representation is improved and there is more stability in trends for all architectures.
{"title":"The Analysis of the Generator Architectures and Loss Functions in Improving the Stability of GANs Training towards Efficient Intrusion Detection","authors":"Raha Soleymanzadeh, R. Kashef","doi":"10.1109/ISCMI56532.2022.10068468","DOIUrl":"https://doi.org/10.1109/ISCMI56532.2022.10068468","url":null,"abstract":"Various research studies have been recently introduced in developing generative models, especially in computer vision and image classification. These models are inspired by a generator and discriminator network architecture in a min-max optimization game called Generative Adversarial Networks (GANs). However, GANs-based models suffer from training instability, which means high oscillations during the training, which provides inaccurate results. There are various causes beyond the instability behaviours, such as the adopted generator architecture, loss function, and distance metrics. In this paper, we focus on the impact of the generator architectures and the loss functions on the GANs training. We aim to provide a comparative assessment of various architectures focusing on ensemble and hybrid models and loss functions such as Focal loss, Binary Cross-Entropy and Mean Squared loss function. Experimental results on NSL-KDD and UNSW-NB15 datasets show that the ensemble models are more stable in terms of training and have higher intrusion detection rates. Additionally, the focal loss can improve the performance of detection minority classes. Using Mean squared loss improved the detection rate for discriminator, however with the Binary Cross entropy loss function, the deep features representation is improved and there is more stability in trends for all architectures.","PeriodicalId":340397,"journal":{"name":"2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)","volume":"162 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126807916","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-11-26DOI: 10.1109/ISCMI56532.2022.10068445
M. Khumalo, K. Prag, K. Nixon
The performance of the Quantum Approximate Optimisation Algorithm (QAOA) in solving the Quadratic Assignment Problem (QAP) is evaluated, with the Variational Quantum Eigensolver (VQE) as a benchmark. The QAP is directly revelant to numerous industry scenarios. The QAP, a Combinatorial Optimisation Problem (COP), is classified as $mathcal{NP}$ -Hard. This classification means CPU time increases exponentially as the problem size scales when solving the QAP using deterministic optimisation techniques. Therefore, this work investigates the QAOA in search of a non-deterministic optimisation technique to efficiently obtain solutions to the QAP. This research compares two warm start techniques to solve QAP instances of sizes 3 to 7. The metrics of comparison - that measure efficiency and solution quality - were introduced in previous work on this topic. For the QAOA, the impact of the p-value, a determination of circuit depth, is investigated. Of the two quantum hybrid heuristics, the VQE retrieves solutions in a shorter computational time with a smaller circuit size, which allows for solving instances with a larger problem size. Compared to the VQE, the QAOA performs better in terms of feasibility as the problem size scales. The quantum warm start method results implies that the QAOA may not maintain higher solution quality for instances larger than size 4. Still, further investigation should be conducted once quantum devices with more qubits and higher quantum volumes are available.
{"title":"Evaluating the performance of the Quantum Approximate Optimisation Algorithm to solve the Quadratic Assignment Problem","authors":"M. Khumalo, K. Prag, K. Nixon","doi":"10.1109/ISCMI56532.2022.10068445","DOIUrl":"https://doi.org/10.1109/ISCMI56532.2022.10068445","url":null,"abstract":"The performance of the Quantum Approximate Optimisation Algorithm (QAOA) in solving the Quadratic Assignment Problem (QAP) is evaluated, with the Variational Quantum Eigensolver (VQE) as a benchmark. The QAP is directly revelant to numerous industry scenarios. The QAP, a Combinatorial Optimisation Problem (COP), is classified as $mathcal{NP}$ -Hard. This classification means CPU time increases exponentially as the problem size scales when solving the QAP using deterministic optimisation techniques. Therefore, this work investigates the QAOA in search of a non-deterministic optimisation technique to efficiently obtain solutions to the QAP. This research compares two warm start techniques to solve QAP instances of sizes 3 to 7. The metrics of comparison - that measure efficiency and solution quality - were introduced in previous work on this topic. For the QAOA, the impact of the p-value, a determination of circuit depth, is investigated. Of the two quantum hybrid heuristics, the VQE retrieves solutions in a shorter computational time with a smaller circuit size, which allows for solving instances with a larger problem size. Compared to the VQE, the QAOA performs better in terms of feasibility as the problem size scales. The quantum warm start method results implies that the QAOA may not maintain higher solution quality for instances larger than size 4. Still, further investigation should be conducted once quantum devices with more qubits and higher quantum volumes are available.","PeriodicalId":340397,"journal":{"name":"2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)","volume":"60 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128013900","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-11-26DOI: 10.1109/ISCMI56532.2022.10068465
Szilvia Jáhn-Erdös, Bence Kövári
Genetic algorithms for NP-complete problems are widespread since it is easy to obtain a solution to the problem. However, its optimality gives the real issue and is not guaranteed to be achievable. In our research, we address a special subproblem of scheduling problems, the final exam scheduling, in which special requirements restrict the state space, which often contradicts each other. The task's difficulty is the massive size of the state space. Genetic algorithm-based solutions were considered since a MILP solver could not find a solution in a reasonable time. A model was built to solve this problem using the genetic algorithm. Most of the possibilities were seen in the different mutation procedures, so we investigated them in more detail. A question for genetic algorithms is what parameters and probabilities to run the model with since the more freedom we give to the run, the larger the runtime. Finding the threshold between the two is essential. Therefore, our experiments measured sizeable real data sets to find the optimal values for this complex problem. The resulting algorithm can significantly facilitate the lengthy manual scheduling processes carried out so far in our university.
{"title":"Exploring the Potential of a Genetic Algorithm on a Real-World Complex Scheduling Problem","authors":"Szilvia Jáhn-Erdös, Bence Kövári","doi":"10.1109/ISCMI56532.2022.10068465","DOIUrl":"https://doi.org/10.1109/ISCMI56532.2022.10068465","url":null,"abstract":"Genetic algorithms for NP-complete problems are widespread since it is easy to obtain a solution to the problem. However, its optimality gives the real issue and is not guaranteed to be achievable. In our research, we address a special subproblem of scheduling problems, the final exam scheduling, in which special requirements restrict the state space, which often contradicts each other. The task's difficulty is the massive size of the state space. Genetic algorithm-based solutions were considered since a MILP solver could not find a solution in a reasonable time. A model was built to solve this problem using the genetic algorithm. Most of the possibilities were seen in the different mutation procedures, so we investigated them in more detail. A question for genetic algorithms is what parameters and probabilities to run the model with since the more freedom we give to the run, the larger the runtime. Finding the threshold between the two is essential. Therefore, our experiments measured sizeable real data sets to find the optimal values for this complex problem. The resulting algorithm can significantly facilitate the lengthy manual scheduling processes carried out so far in our university.","PeriodicalId":340397,"journal":{"name":"2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128649661","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-11-26DOI: 10.1109/ISCMI56532.2022.10068442
Jerome Branny, Rolf Dornberger, T. Hanne
In this paper, we investigate how to forecast Non-Fungible Token (NFT) sale prices by using multiple multivariate time series datasets containing features related to the NFT market space. We examined eight recent studies regarding the forecasting and valuation of NFTs and compared their most important findings. This laid the fundamental work for two separate machine learning prototypes based on Long Short-Term Memory (LSTM) which are able to forecast the sale price history of an individual NFT asset. Root Mean Squared Errors (RMSE) of 0.2975 and 0.24 were obtained which appears to be promising.
{"title":"Non-fungible Token Price Prediction with Multivariate LSTM Neural Networks","authors":"Jerome Branny, Rolf Dornberger, T. Hanne","doi":"10.1109/ISCMI56532.2022.10068442","DOIUrl":"https://doi.org/10.1109/ISCMI56532.2022.10068442","url":null,"abstract":"In this paper, we investigate how to forecast Non-Fungible Token (NFT) sale prices by using multiple multivariate time series datasets containing features related to the NFT market space. We examined eight recent studies regarding the forecasting and valuation of NFTs and compared their most important findings. This laid the fundamental work for two separate machine learning prototypes based on Long Short-Term Memory (LSTM) which are able to forecast the sale price history of an individual NFT asset. Root Mean Squared Errors (RMSE) of 0.2975 and 0.24 were obtained which appears to be promising.","PeriodicalId":340397,"journal":{"name":"2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128681183","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-11-26DOI: 10.1109/ISCMI56532.2022.10068483
T. Dinh, Binh Pham Thanh
Lending activities are an important part of the credit activities of financial institutions and banks. This is an area that brings great potential for development as well as a sustainable source of profit for financial institutions and banks. However, lending to customers also brings high risks. Therefore, predicting the ability to repay on time and understanding the factors affecting the repayment ability of customers is extremely important and necessary, to help financial institutions and banks enhance their ability to pay debts. customers' ability to identify and pay debts on time, contributing to minimizing bad debts and enhancing credit risk management. In this study, Machine Learning models will be used: Proposing a method to combine Logistic Regression with Random Forest, Logistic Regression with K-Nearest Neighbor, Logistic Regression with Support Vector Machine, Logistic Regression with Artificial Neural Network, Logistic Regression with Long short-term memory and finally Logistic Regression with Decision Tree to predict customers' ability to repay on time and compare and evaluate the performance of Machine Learning models. As a result, the Logistic Regression with the Random Forest model ensemble is found as the optimal predictive model and it is expected that Fico Score and annual income significantly influence the forecast.
{"title":"Loan Repayment Prediction Using Logistic Regression Ensemble Learning With Machine Learning Algorithms","authors":"T. Dinh, Binh Pham Thanh","doi":"10.1109/ISCMI56532.2022.10068483","DOIUrl":"https://doi.org/10.1109/ISCMI56532.2022.10068483","url":null,"abstract":"Lending activities are an important part of the credit activities of financial institutions and banks. This is an area that brings great potential for development as well as a sustainable source of profit for financial institutions and banks. However, lending to customers also brings high risks. Therefore, predicting the ability to repay on time and understanding the factors affecting the repayment ability of customers is extremely important and necessary, to help financial institutions and banks enhance their ability to pay debts. customers' ability to identify and pay debts on time, contributing to minimizing bad debts and enhancing credit risk management. In this study, Machine Learning models will be used: Proposing a method to combine Logistic Regression with Random Forest, Logistic Regression with K-Nearest Neighbor, Logistic Regression with Support Vector Machine, Logistic Regression with Artificial Neural Network, Logistic Regression with Long short-term memory and finally Logistic Regression with Decision Tree to predict customers' ability to repay on time and compare and evaluate the performance of Machine Learning models. As a result, the Logistic Regression with the Random Forest model ensemble is found as the optimal predictive model and it is expected that Fico Score and annual income significantly influence the forecast.","PeriodicalId":340397,"journal":{"name":"2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133938416","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-11-26DOI: 10.1109/ISCMI56532.2022.10068459
Méziane Aïder, Imene Dey, M. Hifi
In this paper, we investigate the use of a population-based algorithm for tackling the fuzzy capacitated maximal covering location problem. Such a problem is characterized by a set of customers with their distances and its goal is to determine a subset of locations positioned on customers such that a maximum coverage of customers, including the both fuzzy coverage degree of facilities and the distance between customers, should be optimized. The proposed method is based upon the grey wolf optimizer, which starts by generating an initial population using a greedy rule strategy that is able to achieve feasible solutions according to the current positions of wolves. In order to enhance the quality of solutions induced, a series of local searches are added for exploring the search space by exploiting some nice strategies. The behavior of the method is computationally analyzed on a set of instances of the literature. Encouraging results have been provided.
{"title":"A Population-Based Algorithm for Solving the Fuzzy Capacitated Maximal Covering Location Problem","authors":"Méziane Aïder, Imene Dey, M. Hifi","doi":"10.1109/ISCMI56532.2022.10068459","DOIUrl":"https://doi.org/10.1109/ISCMI56532.2022.10068459","url":null,"abstract":"In this paper, we investigate the use of a population-based algorithm for tackling the fuzzy capacitated maximal covering location problem. Such a problem is characterized by a set of customers with their distances and its goal is to determine a subset of locations positioned on customers such that a maximum coverage of customers, including the both fuzzy coverage degree of facilities and the distance between customers, should be optimized. The proposed method is based upon the grey wolf optimizer, which starts by generating an initial population using a greedy rule strategy that is able to achieve feasible solutions according to the current positions of wolves. In order to enhance the quality of solutions induced, a series of local searches are added for exploring the search space by exploiting some nice strategies. The behavior of the method is computationally analyzed on a set of instances of the literature. Encouraging results have been provided.","PeriodicalId":340397,"journal":{"name":"2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)","volume":"171 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132182983","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-11-26DOI: 10.1109/ISCMI56532.2022.10068484
Jasper Kyle Catapang
In this paper, the author proposes a new transformer model called Hadamard Estimated Attention Transformer or HEAT, that utilizes a low-rank projection of the Hadamard product to approximate the self-attention mechanism in standard transformer architectures and thus aiming to speedup transformer training, finetuning, and inference altogether. The study shows how it is significantly better than the original transformer that uses dot product self-attention by offering a faster way to compute the original self-attention mechanism while maintaining and ultimately surpassing the quality of the original transformer architecture. It also bests Linformer and Nyströmformer in several machine translation tasks while matching and even outperforming Nyströmformer's accuracy in various text classification tasks.
{"title":"Hadamard Estimated Attention Transformer (HEAT): Fast Approximation of Dot Product Self-attention for Transformers Using Low-Rank Projection of Hadamard Product","authors":"Jasper Kyle Catapang","doi":"10.1109/ISCMI56532.2022.10068484","DOIUrl":"https://doi.org/10.1109/ISCMI56532.2022.10068484","url":null,"abstract":"In this paper, the author proposes a new transformer model called Hadamard Estimated Attention Transformer or HEAT, that utilizes a low-rank projection of the Hadamard product to approximate the self-attention mechanism in standard transformer architectures and thus aiming to speedup transformer training, finetuning, and inference altogether. The study shows how it is significantly better than the original transformer that uses dot product self-attention by offering a faster way to compute the original self-attention mechanism while maintaining and ultimately surpassing the quality of the original transformer architecture. It also bests Linformer and Nyströmformer in several machine translation tasks while matching and even outperforming Nyströmformer's accuracy in various text classification tasks.","PeriodicalId":340397,"journal":{"name":"2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)","volume":"316 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134507700","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-11-26DOI: 10.1109/ISCMI56532.2022.10068481
I. Rahimi, Theodore Picard, Andrew Morabito, Kiriakos Pampalis, Aiden Abignano, A. Gandomi
The problem of open-pit mining optimization is a complex task, often containing many variables. In this paper, we apply a trajectory-based algorithm known as simulated annealing together with a well-known population-based algorithm, genetic algorithm, used to generate solutions for a formulation of the constrained pit problem (CPIT). Three datasets were used to test this simulation, Newman1, zuck_small, and KD. The results show that simulated annealing as a trajectory algorithm possesses a slightly better performance in comparison with the genetic algorithm in terms of profit value.
{"title":"Comparison of Trajectory and Population-Based Algorithms for Optimizing Constrained Open-Pit Mining Problem","authors":"I. Rahimi, Theodore Picard, Andrew Morabito, Kiriakos Pampalis, Aiden Abignano, A. Gandomi","doi":"10.1109/ISCMI56532.2022.10068481","DOIUrl":"https://doi.org/10.1109/ISCMI56532.2022.10068481","url":null,"abstract":"The problem of open-pit mining optimization is a complex task, often containing many variables. In this paper, we apply a trajectory-based algorithm known as simulated annealing together with a well-known population-based algorithm, genetic algorithm, used to generate solutions for a formulation of the constrained pit problem (CPIT). Three datasets were used to test this simulation, Newman1, zuck_small, and KD. The results show that simulated annealing as a trajectory algorithm possesses a slightly better performance in comparison with the genetic algorithm in terms of profit value.","PeriodicalId":340397,"journal":{"name":"2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)","volume":"53 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124997707","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-11-26DOI: 10.1109/ISCMI56532.2022.10068463
Abel S. Zacarias, L. Alexandre
Lifelong learning aims to develop machine learning systems that can learn new tasks while preserving the performance on previously learned tasks. Learning new tasks in most proposals, implies to keeping examples of previously learned tasks to retrain the model when learning new tasks, which has an impact in terms of storage capacity. In this paper, we present a method that adds new capabilities, in an incrementally way, to an existing model keeping examples from previously learned classes but avoiding the problem of running out of storage by using distilled images to condensate sets of images into a single image. The experimental results on four data sets confirmed the effectiveness of CILDI to learn new classes incrementally across different tasks and obtaining a performance close to the state-of-the-art algorithms for class incremental learning using only one distilled image per learned class and beating the state-of-the-art on the four data sets when using 10 distilled images per learned class, while using a smaller memory footprint than the competing approaches.
{"title":"CILDI: Class Incremental Learning with Distilled Images","authors":"Abel S. Zacarias, L. Alexandre","doi":"10.1109/ISCMI56532.2022.10068463","DOIUrl":"https://doi.org/10.1109/ISCMI56532.2022.10068463","url":null,"abstract":"Lifelong learning aims to develop machine learning systems that can learn new tasks while preserving the performance on previously learned tasks. Learning new tasks in most proposals, implies to keeping examples of previously learned tasks to retrain the model when learning new tasks, which has an impact in terms of storage capacity. In this paper, we present a method that adds new capabilities, in an incrementally way, to an existing model keeping examples from previously learned classes but avoiding the problem of running out of storage by using distilled images to condensate sets of images into a single image. The experimental results on four data sets confirmed the effectiveness of CILDI to learn new classes incrementally across different tasks and obtaining a performance close to the state-of-the-art algorithms for class incremental learning using only one distilled image per learned class and beating the state-of-the-art on the four data sets when using 10 distilled images per learned class, while using a smaller memory footprint than the competing approaches.","PeriodicalId":340397,"journal":{"name":"2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125309795","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-11-26DOI: 10.1109/ISCMI56532.2022.10068458
Ramona Boeh, T. Hanne, Rolf Dornberger
We compare the two parent selection methods “linear rank” and “tournament” in a Genetic Algorithm applied to a dynamic Travelling Salesman Problem (TSP). The inherent dynamics of the problem is considered by temporarily doubling the costs between two randomly selected cities. In our experiments we take into account tournament selection with tournament sizes of 3, 5, and 10. A larger tournament size results in as good a performance as with linear rank selection in a small-scale dynamic TSP, whereas smaller tournament sizes better preserve the diversity of the population and avoid getting stuck in local optima. However, the assumption that tournament is superior to linear rank on a dynamic TSP could neither be confirmed nor falsified in the applied testcases.
{"title":"A Comparison of Linear Rank and Tournament for Parent Selection in a Genetic Algorithm Solving a Dynamic Travelling Salesman Problem","authors":"Ramona Boeh, T. Hanne, Rolf Dornberger","doi":"10.1109/ISCMI56532.2022.10068458","DOIUrl":"https://doi.org/10.1109/ISCMI56532.2022.10068458","url":null,"abstract":"We compare the two parent selection methods “linear rank” and “tournament” in a Genetic Algorithm applied to a dynamic Travelling Salesman Problem (TSP). The inherent dynamics of the problem is considered by temporarily doubling the costs between two randomly selected cities. In our experiments we take into account tournament selection with tournament sizes of 3, 5, and 10. A larger tournament size results in as good a performance as with linear rank selection in a small-scale dynamic TSP, whereas smaller tournament sizes better preserve the diversity of the population and avoid getting stuck in local optima. However, the assumption that tournament is superior to linear rank on a dynamic TSP could neither be confirmed nor falsified in the applied testcases.","PeriodicalId":340397,"journal":{"name":"2022 9th International Conference on Soft Computing & Machine Intelligence (ISCMI)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123899392","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}