Pub Date : 2020-12-03DOI: 10.2478/jaiscr-2021-0001
J. Rahman, Tom Gedeon, Sabrina Caldwell, Richard Jones, Zi Jin
Abstract Music has the ability to evoke different emotions in people, which is reflected in their physiological signals. Advances in affective computing have introduced computational methods to analyse these signals and understand the relationship between music and emotion in greater detail. We analyse Electrodermal Activity (EDA), Blood Volume Pulse (BVP), Skin Temperature (ST) and Pupil Dilation (PD) collected from 24 participants while they listen to 12 pieces from 3 different genres of music. A set of 34 features were extracted from each signal and 6 different feature selection methods were applied to identify useful features. Empirical analysis shows that a neural network (NN) with a set of features extracted from the physiological signals can achieve 99.2% accuracy in differentiating among the 3 music genres. The model also reaches 98.5% accuracy in classification based on participants’ subjective rating of emotion. The paper also identifies some useful features to improve accuracy of the classification models. Furthermore, we introduce a new technique called ’Gingerbread Animation’ to visualise the physiological signals we record as a video, and to make these signals more comprehensible to the human eye, and also appropriate for computer vision techniques such as Convolutional Neural Networks (CNNs). Our results overall provide a strong motivation to investigate the relationship between physiological signals and music, which can lead to improvements in music therapy for mental health care and musicogenic epilepsy reduction (our long term goal).
{"title":"Towards Effective Music Therapy for Mental Health Care Using Machine Learning Tools: Human Affective Reasoning and Music Genres","authors":"J. Rahman, Tom Gedeon, Sabrina Caldwell, Richard Jones, Zi Jin","doi":"10.2478/jaiscr-2021-0001","DOIUrl":"https://doi.org/10.2478/jaiscr-2021-0001","url":null,"abstract":"Abstract Music has the ability to evoke different emotions in people, which is reflected in their physiological signals. Advances in affective computing have introduced computational methods to analyse these signals and understand the relationship between music and emotion in greater detail. We analyse Electrodermal Activity (EDA), Blood Volume Pulse (BVP), Skin Temperature (ST) and Pupil Dilation (PD) collected from 24 participants while they listen to 12 pieces from 3 different genres of music. A set of 34 features were extracted from each signal and 6 different feature selection methods were applied to identify useful features. Empirical analysis shows that a neural network (NN) with a set of features extracted from the physiological signals can achieve 99.2% accuracy in differentiating among the 3 music genres. The model also reaches 98.5% accuracy in classification based on participants’ subjective rating of emotion. The paper also identifies some useful features to improve accuracy of the classification models. Furthermore, we introduce a new technique called ’Gingerbread Animation’ to visualise the physiological signals we record as a video, and to make these signals more comprehensible to the human eye, and also appropriate for computer vision techniques such as Convolutional Neural Networks (CNNs). Our results overall provide a strong motivation to investigate the relationship between physiological signals and music, which can lead to improvements in music therapy for mental health care and musicogenic epilepsy reduction (our long term goal).","PeriodicalId":48494,"journal":{"name":"Journal of Artificial Intelligence and Soft Computing Research","volume":"11 1","pages":"5 - 20"},"PeriodicalIF":2.8,"publicationDate":"2020-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43287354","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2020-12-03DOI: 10.2478/jaiscr-2021-0003
Julia El Zini, Yara Rizk, M. Awad
Abstract Recurrent neural networks (RNN) have been successfully applied to various sequential decision-making tasks, natural language processing applications, and time-series predictions. Such networks are usually trained through back-propagation through time (BPTT) which is prohibitively expensive, especially when the length of the time dependencies and the number of hidden neurons increase. To reduce the training time, extreme learning machines (ELMs) have been recently applied to RNN training, reaching a 99% speedup on some applications. Due to its non-iterative nature, ELM training, when parallelized, has the potential to reach higher speedups than BPTT. In this work, we present Opt-PR-ELM, an optimized parallel RNN training algorithm based on ELM that takes advantage of the GPU shared memory and of parallel QR factorization algorithms to efficiently reach optimal solutions. The theoretical analysis of the proposed algorithm is presented on six RNN architectures, including LSTM and GRU, and its performance is empirically tested on ten time-series prediction applications. Opt-PR-ELM is shown to reach up to 461 times speedup over its sequential counterpart and to require up to 20x less time to train than parallel BPTT. Such high speedups over new generation CPUs are extremely crucial in real-time applications and IoT environments.
{"title":"An Optimized Parallel Implementation of Non-Iteratively Trained Recurrent Neural Networks","authors":"Julia El Zini, Yara Rizk, M. Awad","doi":"10.2478/jaiscr-2021-0003","DOIUrl":"https://doi.org/10.2478/jaiscr-2021-0003","url":null,"abstract":"Abstract Recurrent neural networks (RNN) have been successfully applied to various sequential decision-making tasks, natural language processing applications, and time-series predictions. Such networks are usually trained through back-propagation through time (BPTT) which is prohibitively expensive, especially when the length of the time dependencies and the number of hidden neurons increase. To reduce the training time, extreme learning machines (ELMs) have been recently applied to RNN training, reaching a 99% speedup on some applications. Due to its non-iterative nature, ELM training, when parallelized, has the potential to reach higher speedups than BPTT. In this work, we present Opt-PR-ELM, an optimized parallel RNN training algorithm based on ELM that takes advantage of the GPU shared memory and of parallel QR factorization algorithms to efficiently reach optimal solutions. The theoretical analysis of the proposed algorithm is presented on six RNN architectures, including LSTM and GRU, and its performance is empirically tested on ten time-series prediction applications. Opt-PR-ELM is shown to reach up to 461 times speedup over its sequential counterpart and to require up to 20x less time to train than parallel BPTT. Such high speedups over new generation CPUs are extremely crucial in real-time applications and IoT environments.","PeriodicalId":48494,"journal":{"name":"Journal of Artificial Intelligence and Soft Computing Research","volume":"11 1","pages":"33 - 50"},"PeriodicalIF":2.8,"publicationDate":"2020-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47022911","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2020-06-15DOI: 10.2478/jaiscr-2020-0017
Xuan Zhao, Meichen Song, Anqi Liu, Yiming Wang, Tong Wang, Jinde Cao
Abstract Air quality data prediction in urban area is of great significance to control air pollution and protect the public health. The prediction of the air quality in the monitoring station is well studied in existing researches. However, air-quality-monitor stations are insufficient in most cities and the air quality varies from one place to another dramatically due to complex factors. A novel model is established in this paper to estimate and predict the Air Quality Index (AQI) of the areas without monitoring stations in Nanjing. The proposed model predicts AQI in a non-monitoring area both in temporal dimension and in spatial dimension respectively. The temporal dimension model is presented at first based on the enhanced k-Nearest Neighbor (KNN) algorithm to predict the AQI values among monitoring stations, the acceptability of the results achieves 92% for one-hour prediction. Meanwhile, in order to forecast the evolution of air quality in the spatial dimension, the method is utilized with the help of Back Propagation neural network (BP), which considers geographical distance. Furthermore, to improve the accuracy and adaptability of the spatial model, the similarity of topological structure is introduced. Especially, the temporal-spatial model is built and its adaptability is tested on a specific non-monitoring site, Jiulonghu Campus of Southeast University. The result demonstrates that the acceptability achieves 73.8% on average. The current paper provides strong evidence suggesting that the proposed non-parametric and data-driven approach for air quality forecasting provides promising results.
{"title":"Data-Driven Temporal-Spatial Model for the Prediction of AQI in Nanjing","authors":"Xuan Zhao, Meichen Song, Anqi Liu, Yiming Wang, Tong Wang, Jinde Cao","doi":"10.2478/jaiscr-2020-0017","DOIUrl":"https://doi.org/10.2478/jaiscr-2020-0017","url":null,"abstract":"Abstract Air quality data prediction in urban area is of great significance to control air pollution and protect the public health. The prediction of the air quality in the monitoring station is well studied in existing researches. However, air-quality-monitor stations are insufficient in most cities and the air quality varies from one place to another dramatically due to complex factors. A novel model is established in this paper to estimate and predict the Air Quality Index (AQI) of the areas without monitoring stations in Nanjing. The proposed model predicts AQI in a non-monitoring area both in temporal dimension and in spatial dimension respectively. The temporal dimension model is presented at first based on the enhanced k-Nearest Neighbor (KNN) algorithm to predict the AQI values among monitoring stations, the acceptability of the results achieves 92% for one-hour prediction. Meanwhile, in order to forecast the evolution of air quality in the spatial dimension, the method is utilized with the help of Back Propagation neural network (BP), which considers geographical distance. Furthermore, to improve the accuracy and adaptability of the spatial model, the similarity of topological structure is introduced. Especially, the temporal-spatial model is built and its adaptability is tested on a specific non-monitoring site, Jiulonghu Campus of Southeast University. The result demonstrates that the acceptability achieves 73.8% on average. The current paper provides strong evidence suggesting that the proposed non-parametric and data-driven approach for air quality forecasting provides promising results.","PeriodicalId":48494,"journal":{"name":"Journal of Artificial Intelligence and Soft Computing Research","volume":"10 1","pages":"255 - 270"},"PeriodicalIF":2.8,"publicationDate":"2020-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49397652","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2020-06-15DOI: 10.2478/jaiscr-2020-0016
Marcin Gabryel, K. Grzanek, Y. Hayashi
Abstract Web-based browser fingerprint (or device fingerprint) is a tool used to identify and track user activity in web traffic. It is also used to identify computers that are abusing online advertising and also to prevent credit card fraud. A device fingerprint is created by extracting multiple parameter values from a browser API (e.g. operating system type or browser version). The acquired parameter values are then used to create a hash using the hash function. The disadvantage of using this method is too high susceptibility to small, normally occurring changes (e.g. when changing the browser version number or screen resolution). Minor changes in the input values generate a completely different fingerprint hash, making it impossible to find similar ones in the database. On the other hand, omitting these unstable values when creating a hash, significantly limits the ability of the fingerprint to distinguish between devices. This weak point is commonly exploited by fraudsters who knowingly evade this form of protection by deliberately changing the value of device parameters. The paper presents methods that significantly limit this type of activity. New algorithms for coding and comparing fingerprints are presented, in which the values of parameters with low stability and low entropy are especially taken into account. The fingerprint generation methods are based on popular Minhash, the LSH, and autoencoder methods. The effectiveness of coding and comparing each of the presented methods was also examined in comparison with the currently used hash generation method. Authentic data of the devices and browsers of users visiting 186 different websites were collected for the research.
{"title":"Browser Fingerprint Coding Methods Increasing the Effectiveness of User Identification in the Web Traffic","authors":"Marcin Gabryel, K. Grzanek, Y. Hayashi","doi":"10.2478/jaiscr-2020-0016","DOIUrl":"https://doi.org/10.2478/jaiscr-2020-0016","url":null,"abstract":"Abstract Web-based browser fingerprint (or device fingerprint) is a tool used to identify and track user activity in web traffic. It is also used to identify computers that are abusing online advertising and also to prevent credit card fraud. A device fingerprint is created by extracting multiple parameter values from a browser API (e.g. operating system type or browser version). The acquired parameter values are then used to create a hash using the hash function. The disadvantage of using this method is too high susceptibility to small, normally occurring changes (e.g. when changing the browser version number or screen resolution). Minor changes in the input values generate a completely different fingerprint hash, making it impossible to find similar ones in the database. On the other hand, omitting these unstable values when creating a hash, significantly limits the ability of the fingerprint to distinguish between devices. This weak point is commonly exploited by fraudsters who knowingly evade this form of protection by deliberately changing the value of device parameters. The paper presents methods that significantly limit this type of activity. New algorithms for coding and comparing fingerprints are presented, in which the values of parameters with low stability and low entropy are especially taken into account. The fingerprint generation methods are based on popular Minhash, the LSH, and autoencoder methods. The effectiveness of coding and comparing each of the presented methods was also examined in comparison with the currently used hash generation method. Authentic data of the devices and browsers of users visiting 186 different websites were collected for the research.","PeriodicalId":48494,"journal":{"name":"Journal of Artificial Intelligence and Soft Computing Research","volume":"10 1","pages":"243 - 253"},"PeriodicalIF":2.8,"publicationDate":"2020-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45328315","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2020-06-15DOI: 10.2478/jaiscr-2020-0020
J. Bilski, Bartosz Kowalczyk, A. Marchlewska, J. Zurada
Abstract This paper presents a local modification of the Levenberg-Marquardt algorithm (LM). First, the mathematical basics of the classic LM method are shown. The classic LM algorithm is very efficient for learning small neural networks. For bigger neural networks, whose computational complexity grows significantly, it makes this method practically inefficient. In order to overcome this limitation, local modification of the LM is introduced in this paper. The main goal of this paper is to develop a more complexity efficient modification of the LM method by using a local computation. The introduced modification has been tested on the following benchmarks: the function approximation and classification problems. The obtained results have been compared to the classic LM method performance. The paper shows that the local modification of the LM method significantly improves the algorithm’s performance for bigger networks. Several possible proposals for future works are suggested.
{"title":"Local Levenberg-Marquardt Algorithm for Learning Feedforwad Neural Networks","authors":"J. Bilski, Bartosz Kowalczyk, A. Marchlewska, J. Zurada","doi":"10.2478/jaiscr-2020-0020","DOIUrl":"https://doi.org/10.2478/jaiscr-2020-0020","url":null,"abstract":"Abstract This paper presents a local modification of the Levenberg-Marquardt algorithm (LM). First, the mathematical basics of the classic LM method are shown. The classic LM algorithm is very efficient for learning small neural networks. For bigger neural networks, whose computational complexity grows significantly, it makes this method practically inefficient. In order to overcome this limitation, local modification of the LM is introduced in this paper. The main goal of this paper is to develop a more complexity efficient modification of the LM method by using a local computation. The introduced modification has been tested on the following benchmarks: the function approximation and classification problems. The obtained results have been compared to the classic LM method performance. The paper shows that the local modification of the LM method significantly improves the algorithm’s performance for bigger networks. Several possible proposals for future works are suggested.","PeriodicalId":48494,"journal":{"name":"Journal of Artificial Intelligence and Soft Computing Research","volume":"10 1","pages":"299 - 316"},"PeriodicalIF":2.8,"publicationDate":"2020-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48966110","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2020-06-15DOI: 10.2478/jaiscr-2020-0019
P. Duda, K. Przybyszewski, Lipo Wang
Abstract The training set consists of many features that influence the classifier in different degrees. Choosing the most important features and rejecting those that do not carry relevant information is of great importance to the operating of the learned model. In the case of data streams, the importance of the features may additionally change over time. Such changes affect the performance of the classifier but can also be an important indicator of occurring concept-drift. In this work, we propose a new algorithm for data streams classification, called Random Forest with Features Importance (RFFI), which uses the measure of features importance as a drift detector. The RFFT algorithm implements solutions inspired by the Random Forest algorithm to the data stream scenarios. The proposed algorithm combines the ability of ensemble methods for handling slow changes in a data stream with a new method for detecting concept drift occurrence. The work contains an experimental analysis of the proposed algorithm, carried out on synthetic and real data.
{"title":"A Novel Drift Detection Algorithm Based on Features’ Importance Analysis in a Data Streams Environment","authors":"P. Duda, K. Przybyszewski, Lipo Wang","doi":"10.2478/jaiscr-2020-0019","DOIUrl":"https://doi.org/10.2478/jaiscr-2020-0019","url":null,"abstract":"Abstract The training set consists of many features that influence the classifier in different degrees. Choosing the most important features and rejecting those that do not carry relevant information is of great importance to the operating of the learned model. In the case of data streams, the importance of the features may additionally change over time. Such changes affect the performance of the classifier but can also be an important indicator of occurring concept-drift. In this work, we propose a new algorithm for data streams classification, called Random Forest with Features Importance (RFFI), which uses the measure of features importance as a drift detector. The RFFT algorithm implements solutions inspired by the Random Forest algorithm to the data stream scenarios. The proposed algorithm combines the ability of ensemble methods for handling slow changes in a data stream with a new method for detecting concept drift occurrence. The work contains an experimental analysis of the proposed algorithm, carried out on synthetic and real data.","PeriodicalId":48494,"journal":{"name":"Journal of Artificial Intelligence and Soft Computing Research","volume":"10 1","pages":"287 - 298"},"PeriodicalIF":2.8,"publicationDate":"2020-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41853974","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2020-06-15DOI: 10.2478/jaiscr-2020-0018
Janusz T. Starczewski, P. Goetzen, Christian Napoli
Abstract In real-world approximation problems, precise input data are economically expensive. Therefore, fuzzy methods devoted to uncertain data are in the focus of current research. Consequently, a method based on fuzzy-rough sets for fuzzification of inputs in a rule-based fuzzy system is discussed in this paper. A triangular membership function is applied to describe the nature of imprecision in data. Firstly, triangular fuzzy partitions are introduced to approximate common antecedent fuzzy rule sets. As a consequence of the proposed method, we obtain a structure of a general (non-interval) type-2 fuzzy logic system in which secondary membership functions are cropped triangular. Then, the possibility of applying so-called regular triangular norms is discussed. Finally, an experimental system constructed on precise data, which is then transformed and verified for uncertain data, is provided to demonstrate its basic properties.
{"title":"Triangular Fuzzy-Rough Set Based Fuzzification of Fuzzy Rule-Based Systems","authors":"Janusz T. Starczewski, P. Goetzen, Christian Napoli","doi":"10.2478/jaiscr-2020-0018","DOIUrl":"https://doi.org/10.2478/jaiscr-2020-0018","url":null,"abstract":"Abstract In real-world approximation problems, precise input data are economically expensive. Therefore, fuzzy methods devoted to uncertain data are in the focus of current research. Consequently, a method based on fuzzy-rough sets for fuzzification of inputs in a rule-based fuzzy system is discussed in this paper. A triangular membership function is applied to describe the nature of imprecision in data. Firstly, triangular fuzzy partitions are introduced to approximate common antecedent fuzzy rule sets. As a consequence of the proposed method, we obtain a structure of a general (non-interval) type-2 fuzzy logic system in which secondary membership functions are cropped triangular. Then, the possibility of applying so-called regular triangular norms is discussed. Finally, an experimental system constructed on precise data, which is then transformed and verified for uncertain data, is provided to demonstrate its basic properties.","PeriodicalId":48494,"journal":{"name":"Journal of Artificial Intelligence and Soft Computing Research","volume":"10 1","pages":"271 - 285"},"PeriodicalIF":2.8,"publicationDate":"2020-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42745671","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2020-05-23DOI: 10.2478/jaiscr-2020-0012
Marcin Zalasiński, K. Cpałka, Lukasz Laskowski, D. Wunsch, K. Przybyszewski
Abstract In biometrics, methods which are able to precisely adapt to the biometric features of users are much sought after. They use various methods of artificial intelligence, in particular methods from the group of soft computing. In this paper, we focus on on-line signature verification. Such signatures are complex objects described not only by the shape but also by the dynamics of the signing process. In standard devices used for signature acquisition (with an LCD touch screen) this dynamics may include pen velocity, but sometimes other types of signals are also available, e.g. pen pressure on the screen surface (e.g. in graphic tablets), the angle between the pen and the screen surface, etc. The precision of the on-line signature dynamics processing has been a motivational springboard for developing methods that use signature partitioning. Partitioning uses a well-known principle of decomposing the problem into smaller ones. In this paper, we propose a new partitioning algorithm that uses capabilities of the algorithms based on populations and fuzzy systems. Evolutionary-fuzzy partitioning eliminates the need to average dynamic waveforms in created partitions because it replaces them. Evolutionary separation of partitions results in a better matching of partitions with reference signatures, eliminates dispro-portions between the number of points describing dynamics in partitions, eliminates the impact of random values, separates partitions related to the signing stage and its dynamics (e.g. high and low velocity of signing, where high and low are imprecise-fuzzy concepts). The operation of the presented algorithm has been tested using the well-known BioSecure DS2 database of real dynamic signatures.
{"title":"An Algorithm for the Evolutionary-Fuzzy Generation of on-Line Signature Hybrid Descriptors","authors":"Marcin Zalasiński, K. Cpałka, Lukasz Laskowski, D. Wunsch, K. Przybyszewski","doi":"10.2478/jaiscr-2020-0012","DOIUrl":"https://doi.org/10.2478/jaiscr-2020-0012","url":null,"abstract":"Abstract In biometrics, methods which are able to precisely adapt to the biometric features of users are much sought after. They use various methods of artificial intelligence, in particular methods from the group of soft computing. In this paper, we focus on on-line signature verification. Such signatures are complex objects described not only by the shape but also by the dynamics of the signing process. In standard devices used for signature acquisition (with an LCD touch screen) this dynamics may include pen velocity, but sometimes other types of signals are also available, e.g. pen pressure on the screen surface (e.g. in graphic tablets), the angle between the pen and the screen surface, etc. The precision of the on-line signature dynamics processing has been a motivational springboard for developing methods that use signature partitioning. Partitioning uses a well-known principle of decomposing the problem into smaller ones. In this paper, we propose a new partitioning algorithm that uses capabilities of the algorithms based on populations and fuzzy systems. Evolutionary-fuzzy partitioning eliminates the need to average dynamic waveforms in created partitions because it replaces them. Evolutionary separation of partitions results in a better matching of partitions with reference signatures, eliminates dispro-portions between the number of points describing dynamics in partitions, eliminates the impact of random values, separates partitions related to the signing stage and its dynamics (e.g. high and low velocity of signing, where high and low are imprecise-fuzzy concepts). The operation of the presented algorithm has been tested using the well-known BioSecure DS2 database of real dynamic signatures.","PeriodicalId":48494,"journal":{"name":"Journal of Artificial Intelligence and Soft Computing Research","volume":"10 1","pages":"173 - 187"},"PeriodicalIF":2.8,"publicationDate":"2020-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44718088","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2020-05-23DOI: 10.2478/jaiscr-2020-0014
Artur Starczewski, P. Goetzen, M. Er
Abstract Clustering is an attractive technique used in many fields in order to deal with large scale data. Many clustering algorithms have been proposed so far. The most popular algorithms include density-based approaches. These kinds of algorithms can identify clusters of arbitrary shapes in datasets. The most common of them is the Density-Based Spatial Clustering of Applications with Noise (DBSCAN). The original DBSCAN algorithm has been widely applied in various applications and has many different modifications. However, there is a fundamental issue of the right choice of its two input parameters, i.e the eps radius and the MinPts density threshold. The choice of these parameters is especially difficult when the density variation within clusters is significant. In this paper, a new method that determines the right values of the parameters for different kinds of clusters is proposed. This method uses detection of sharp distance increases generated by a function which computes a distance between each element of a dataset and its k-th nearest neighbor. Experimental results have been obtained for several different datasets and they confirm a very good performance of the newly proposed method.
{"title":"A New Method for Automatic Determining of the DBSCAN Parameters","authors":"Artur Starczewski, P. Goetzen, M. Er","doi":"10.2478/jaiscr-2020-0014","DOIUrl":"https://doi.org/10.2478/jaiscr-2020-0014","url":null,"abstract":"Abstract Clustering is an attractive technique used in many fields in order to deal with large scale data. Many clustering algorithms have been proposed so far. The most popular algorithms include density-based approaches. These kinds of algorithms can identify clusters of arbitrary shapes in datasets. The most common of them is the Density-Based Spatial Clustering of Applications with Noise (DBSCAN). The original DBSCAN algorithm has been widely applied in various applications and has many different modifications. However, there is a fundamental issue of the right choice of its two input parameters, i.e the eps radius and the MinPts density threshold. The choice of these parameters is especially difficult when the density variation within clusters is significant. In this paper, a new method that determines the right values of the parameters for different kinds of clusters is proposed. This method uses detection of sharp distance increases generated by a function which computes a distance between each element of a dataset and its k-th nearest neighbor. Experimental results have been obtained for several different datasets and they confirm a very good performance of the newly proposed method.","PeriodicalId":48494,"journal":{"name":"Journal of Artificial Intelligence and Soft Computing Research","volume":"10 1","pages":"209 - 221"},"PeriodicalIF":2.8,"publicationDate":"2020-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47729037","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2020-05-23DOI: 10.2478/jaiscr-2020-0015
Rafał Grycuk, A. Wojciechowski, Wei Wei, A. Siwocha
Abstract Content-based image retrieval methods develop rapidly with a growing scale of image repositories. They are usually based on comparing and indexing some image features. We developed a new algorithm for finding objects in images by traversing their edges. Moreover, we describe the objects by histograms of local features and angles. We use such a description to retrieve similar images fast. We performed extensive experiments on three established image datasets proving the effectiveness of the proposed method.
{"title":"Detecting Visual Objects by Edge Crawling","authors":"Rafał Grycuk, A. Wojciechowski, Wei Wei, A. Siwocha","doi":"10.2478/jaiscr-2020-0015","DOIUrl":"https://doi.org/10.2478/jaiscr-2020-0015","url":null,"abstract":"Abstract Content-based image retrieval methods develop rapidly with a growing scale of image repositories. They are usually based on comparing and indexing some image features. We developed a new algorithm for finding objects in images by traversing their edges. Moreover, we describe the objects by histograms of local features and angles. We use such a description to retrieve similar images fast. We performed extensive experiments on three established image datasets proving the effectiveness of the proposed method.","PeriodicalId":48494,"journal":{"name":"Journal of Artificial Intelligence and Soft Computing Research","volume":"10 1","pages":"223 - 237"},"PeriodicalIF":2.8,"publicationDate":"2020-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47958028","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}