首页 > 最新文献

Journal of Artificial Intelligence and Soft Computing Research最新文献

英文 中文
Classification of Unwanted E-Mails (Spam) with Turkish Text by Different Algorithms in Weka Program 用不同的算法在Weka程序中分类不需要的电子邮件(垃圾邮件)与土耳其文本
IF 2.8 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2022-05-01 DOI: 10.55195/jscai.1104694
H. Simsek, Emrah Aydemir
{"title":"Classification of Unwanted E-Mails (Spam) with Turkish Text by Different Algorithms in Weka Program","authors":"H. Simsek, Emrah Aydemir","doi":"10.55195/jscai.1104694","DOIUrl":"https://doi.org/10.55195/jscai.1104694","url":null,"abstract":"","PeriodicalId":48494,"journal":{"name":"Journal of Artificial Intelligence and Soft Computing Research","volume":"18 1","pages":""},"PeriodicalIF":2.8,"publicationDate":"2022-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89611469","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Anomaly Pattern Detection in Streaming Data Based on the Transformation to Multiple Binary-Valued Data Streams 基于向多个二进制值数据流转换的流数据异常模式检测
IF 2.8 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2021-10-08 DOI: 10.2478/jaiscr-2022-0002
Taegong Kim, C. Park
Abstract Anomaly pattern detection in a data stream aims to detect a time point where outliers begin to occur abnormally. Recently, a method for anomaly pattern detection has been proposed based on binary classification for outliers and statistical tests in the data stream of binary labels of normal or an outlier. It showed that an anomaly pattern can be detected accurately even when outlier detection performance is relatively low. However, since the anomaly pattern detection method is based on the binary classification for outliers, most well-known outlier detection methods, with the output of real-valued outlier scores, can not be used directly. In this paper, we propose an anomaly pattern detection method in a data stream using the transformation to multiple binary-valued data streams from real-valued outlier scores. By using three outlier detection methods, Isolation Forest(IF), Autoencoder-based outlier detection, and Local outlier factor(LOF), the proposed anomaly pattern detection method is tested using artificial and real data sets. The experimental results show that anomaly pattern detection using Isolation Forest gives the best performance.
摘要数据流中的异常模式检测旨在检测异常值开始异常出现的时间点。最近,提出了一种基于异常值的二进制分类和正常或异常值的二值标签数据流中的统计检验的异常模式检测方法。结果表明,即使在异常点检测性能相对较低的情况下,也可以准确地检测到异常模式。然而,由于异常模式检测方法是基于对异常值的二元分类,因此大多数已知的异常值检测方法都不能直接使用,其输出的是实值异常值分数。在本文中,我们提出了一种数据流中的异常模式检测方法,该方法使用从实值异常值分数到多个二进制值数据流的转换。通过使用隔离森林(IF)、基于自动编码器的异常值检测和局部异常值因子(LOF)三种异常值检测方法,使用人工和真实数据集对所提出的异常模式检测方法进行了测试。实验结果表明,使用隔离森林的异常模式检测具有最好的性能。
{"title":"Anomaly Pattern Detection in Streaming Data Based on the Transformation to Multiple Binary-Valued Data Streams","authors":"Taegong Kim, C. Park","doi":"10.2478/jaiscr-2022-0002","DOIUrl":"https://doi.org/10.2478/jaiscr-2022-0002","url":null,"abstract":"Abstract Anomaly pattern detection in a data stream aims to detect a time point where outliers begin to occur abnormally. Recently, a method for anomaly pattern detection has been proposed based on binary classification for outliers and statistical tests in the data stream of binary labels of normal or an outlier. It showed that an anomaly pattern can be detected accurately even when outlier detection performance is relatively low. However, since the anomaly pattern detection method is based on the binary classification for outliers, most well-known outlier detection methods, with the output of real-valued outlier scores, can not be used directly. In this paper, we propose an anomaly pattern detection method in a data stream using the transformation to multiple binary-valued data streams from real-valued outlier scores. By using three outlier detection methods, Isolation Forest(IF), Autoencoder-based outlier detection, and Local outlier factor(LOF), the proposed anomaly pattern detection method is tested using artificial and real data sets. The experimental results show that anomaly pattern detection using Isolation Forest gives the best performance.","PeriodicalId":48494,"journal":{"name":"Journal of Artificial Intelligence and Soft Computing Research","volume":"12 1","pages":"19 - 27"},"PeriodicalIF":2.8,"publicationDate":"2021-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44011186","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A New Hand-Movement-Based Authentication Method Using Feature Importance Selection with the Hotelling’s Statistic 基于Hotelling统计特征重要性选择的手部动作认证新方法
IF 2.8 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2021-10-08 DOI: 10.2478/jaiscr-2022-0004
R. Doroz, K. Wrobel, P. Porwik, T. Orczyk
Abstract The growing amount of collected and processed data means that there is a need to control access to these resources. Very often, this type of control is carried out on the basis of bio-metric analysis. The article proposes a new user authentication method based on a spatial analysis of the movement of the finger’s position. This movement creates a sequence of data that is registered by a motion recording device. The presented approach combines spatial analysis of the position of all fingers at the time. The proposed method is able to use the specific, often different movements of fingers of each user. The experimental results confirm the effectiveness of the method in biometric applications. In this paper, we also introduce an effective method of feature selection, based on the Hotelling T2 statistic. This approach allows selecting the best distinctive features of each object from a set of all objects in the database. It is possible thanks to the appropriate preparation of the input data.
收集和处理的数据量越来越大,这意味着需要控制对这些资源的访问。通常,这种类型的控制是在生物识别分析的基础上进行的。本文提出了一种新的基于手指位置运动空间分析的用户认证方法。这个动作产生一系列的数据,这些数据被动作记录设备记录下来。所提出的方法结合了所有手指在同一时间的位置空间分析。所提出的方法能够使用每个用户特定的,通常不同的手指运动。实验结果证实了该方法在生物识别应用中的有效性。本文还介绍了一种有效的基于Hotelling T2统计量的特征选择方法。这种方法允许从数据库中所有对象的集合中选择每个对象的最佳独特特征。由于输入数据的适当准备,这是可能的。
{"title":"A New Hand-Movement-Based Authentication Method Using Feature Importance Selection with the Hotelling’s Statistic","authors":"R. Doroz, K. Wrobel, P. Porwik, T. Orczyk","doi":"10.2478/jaiscr-2022-0004","DOIUrl":"https://doi.org/10.2478/jaiscr-2022-0004","url":null,"abstract":"Abstract The growing amount of collected and processed data means that there is a need to control access to these resources. Very often, this type of control is carried out on the basis of bio-metric analysis. The article proposes a new user authentication method based on a spatial analysis of the movement of the finger’s position. This movement creates a sequence of data that is registered by a motion recording device. The presented approach combines spatial analysis of the position of all fingers at the time. The proposed method is able to use the specific, often different movements of fingers of each user. The experimental results confirm the effectiveness of the method in biometric applications. In this paper, we also introduce an effective method of feature selection, based on the Hotelling T2 statistic. This approach allows selecting the best distinctive features of each object from a set of all objects in the database. It is possible thanks to the appropriate preparation of the input data.","PeriodicalId":48494,"journal":{"name":"Journal of Artificial Intelligence and Soft Computing Research","volume":"12 1","pages":"41 - 59"},"PeriodicalIF":2.8,"publicationDate":"2021-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43889186","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Evaluating Dropout Placements in Bayesian Regression Resnet 在贝叶斯回归Resnet中评估辍学安置
IF 2.8 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2021-10-08 DOI: 10.2478/jaiscr-2022-0005
Lei Shi, C. Copot, S. Vanlanduit
Abstract Deep Neural Networks (DNNs) have shown great success in many fields. Various network architectures have been developed for different applications. Regardless of the complexities of the networks, DNNs do not provide model uncertainty. Bayesian Neural Networks (BNNs), on the other hand, is able to make probabilistic inference. Among various types of BNNs, Dropout as a Bayesian Approximation converts a Neural Network (NN) to a BNN by adding a dropout layer after each weight layer in the NN. This technique provides a simple transformation from a NN to a BNN. However, for DNNs, adding a dropout layer to each weight layer would lead to a strong regularization due to the deep architecture. Previous researches [1, 2, 3] have shown that adding a dropout layer after each weight layer in a DNN is unnecessary. However, how to place dropout layers in a ResNet for regression tasks are less explored. In this work, we perform an empirical study on how different dropout placements would affect the performance of a Bayesian DNN. We use a regression model modified from ResNet as the DNN and place the dropout layers at different places in the regression ResNet. Our experimental results show that it is not necessary to add a dropout layer after every weight layer in the Regression ResNet to let it be able to make Bayesian Inference. Placing Dropout layers between the stacked blocks i.e. Dense+Identity+Identity blocks has the best performance in Predictive Interval Coverage Probability (PICP). Placing a dropout layer after each stacked block has the best performance in Root Mean Square Error (RMSE).
摘要深度神经网络(Deep Neural Networks, dnn)在许多领域都取得了巨大的成功。针对不同的应用开发了不同的网络体系结构。不管网络的复杂性如何,深度神经网络不提供模型不确定性。另一方面,贝叶斯神经网络(BNNs)能够进行概率推理。在各种类型的BNN中,Dropout作为一种贝叶斯近似,通过在神经网络(NN)的每个权重层之后添加Dropout层,将神经网络(NN)转换为BNN。该技术提供了从NN到BNN的简单转换。然而,对于dnn来说,在每个权重层上添加一个dropout层将导致由于深度架构而产生强正则化。先前的研究[1,2,3]表明,在DNN的每个权重层之后添加dropout层是不必要的。然而,如何在ResNet中放置dropout层进行回归任务的探索较少。在这项工作中,我们对不同的退出位置如何影响贝叶斯深度神经网络的性能进行了实证研究。我们使用从ResNet修改的回归模型作为深度神经网络,并将dropout层放置在回归ResNet的不同位置。我们的实验结果表明,在Regression ResNet中,不需要在每个权重层之后添加dropout层来使其能够进行贝叶斯推理。将Dropout层放置在堆叠块(即Dense+Identity+Identity块)之间具有最佳的预测间隔覆盖概率(PICP)性能。在每个堆叠块之后放置一个dropout层,在均方根误差(RMSE)方面具有最佳性能。
{"title":"Evaluating Dropout Placements in Bayesian Regression Resnet","authors":"Lei Shi, C. Copot, S. Vanlanduit","doi":"10.2478/jaiscr-2022-0005","DOIUrl":"https://doi.org/10.2478/jaiscr-2022-0005","url":null,"abstract":"Abstract Deep Neural Networks (DNNs) have shown great success in many fields. Various network architectures have been developed for different applications. Regardless of the complexities of the networks, DNNs do not provide model uncertainty. Bayesian Neural Networks (BNNs), on the other hand, is able to make probabilistic inference. Among various types of BNNs, Dropout as a Bayesian Approximation converts a Neural Network (NN) to a BNN by adding a dropout layer after each weight layer in the NN. This technique provides a simple transformation from a NN to a BNN. However, for DNNs, adding a dropout layer to each weight layer would lead to a strong regularization due to the deep architecture. Previous researches [1, 2, 3] have shown that adding a dropout layer after each weight layer in a DNN is unnecessary. However, how to place dropout layers in a ResNet for regression tasks are less explored. In this work, we perform an empirical study on how different dropout placements would affect the performance of a Bayesian DNN. We use a regression model modified from ResNet as the DNN and place the dropout layers at different places in the regression ResNet. Our experimental results show that it is not necessary to add a dropout layer after every weight layer in the Regression ResNet to let it be able to make Bayesian Inference. Placing Dropout layers between the stacked blocks i.e. Dense+Identity+Identity blocks has the best performance in Predictive Interval Coverage Probability (PICP). Placing a dropout layer after each stacked block has the best performance in Root Mean Square Error (RMSE).","PeriodicalId":48494,"journal":{"name":"Journal of Artificial Intelligence and Soft Computing Research","volume":"12 1","pages":"61 - 73"},"PeriodicalIF":2.8,"publicationDate":"2021-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48117522","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Performance Analysis of Data Fusion Methods Applied to Epileptic Seizure Recognition 癫痫发作识别中数据融合方法的性能分析
IF 2.8 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2021-10-08 DOI: 10.2478/jaiscr-2022-0001
Simone A. Ludwig
Abstract Epilepsy is a chronic neurological disorder that is caused by unprovoked recurrent seizures. The most commonly used tool for the diagnosis of epilepsy is the electroencephalogram (EEG) whereby the electrical activity of the brain is measured. In order to prevent potential risks, the patients have to be monitored as to detect an epileptic episode early on and to provide prevention measures. Many different research studies have used a combination of time and frequency features for the automatic recognition of epileptic seizures. In this paper, two fusion methods are compared. The first is based on an ensemble method and the second uses the Choquet fuzzy integral method. In particular, three different machine learning approaches namely RNN, ML and DNN are used as inputs for the ensemble method and the Choquet fuzzy integral fusion method. Evaluation measures such as confusion matrix, AUC and accuracy are compared as well as MSE and RMSE are provided. The results show that the Choquet fuzzy integral fusion method outperforms the ensemble method as well as other state-of-the-art classification methods.
摘要癫痫是一种由无端反复发作引起的慢性神经系统疾病。诊断癫痫最常用的工具是脑电图(EEG),通过脑电图可以测量大脑的电活动。为了防止潜在的风险,必须对患者进行监测,以便尽早发现癫痫发作并提供预防措施。许多不同的研究都使用了时间和频率特征的组合来自动识别癫痫发作。本文对两种融合方法进行了比较。第一种方法基于集成方法,第二种方法使用Choquet模糊积分方法。特别地,三种不同的机器学习方法,即RNN、ML和DNN,被用作集成方法和Choquet模糊积分融合方法的输入。比较了混淆矩阵、AUC和准确性等评估指标,并提供了MSE和RMSE。结果表明,Choquet模糊积分融合方法优于集成方法以及其他先进的分类方法。
{"title":"Performance Analysis of Data Fusion Methods Applied to Epileptic Seizure Recognition","authors":"Simone A. Ludwig","doi":"10.2478/jaiscr-2022-0001","DOIUrl":"https://doi.org/10.2478/jaiscr-2022-0001","url":null,"abstract":"Abstract Epilepsy is a chronic neurological disorder that is caused by unprovoked recurrent seizures. The most commonly used tool for the diagnosis of epilepsy is the electroencephalogram (EEG) whereby the electrical activity of the brain is measured. In order to prevent potential risks, the patients have to be monitored as to detect an epileptic episode early on and to provide prevention measures. Many different research studies have used a combination of time and frequency features for the automatic recognition of epileptic seizures. In this paper, two fusion methods are compared. The first is based on an ensemble method and the second uses the Choquet fuzzy integral method. In particular, three different machine learning approaches namely RNN, ML and DNN are used as inputs for the ensemble method and the Choquet fuzzy integral fusion method. Evaluation measures such as confusion matrix, AUC and accuracy are compared as well as MSE and RMSE are provided. The results show that the Choquet fuzzy integral fusion method outperforms the ensemble method as well as other state-of-the-art classification methods.","PeriodicalId":48494,"journal":{"name":"Journal of Artificial Intelligence and Soft Computing Research","volume":"12 1","pages":"5 - 17"},"PeriodicalIF":2.8,"publicationDate":"2021-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49667037","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Mixup (Sample Pairing) Can Improve the Performance of Deep Segmentation Networks 混合(样本配对)可以提高深度分割网络的性能
IF 2.8 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2021-10-08 DOI: 10.2478/jaiscr-2022-0003
L. Isaksson, P. Summers, S. Raimondi, S. Gandini, A. Bhalerao, G. Marvaso, G. Petralia, M. Pepa, B. Jereczek-Fossa
Abstract Researchers address the generalization problem of deep image processing networks mainly through extensive use of data augmentation techniques such as random flips, rotations, and deformations. A data augmentation technique called mixup, which constructs virtual training samples from convex combinations of inputs, was recently proposed for deep classification networks. The algorithm contributed to increased performance on classification in a variety of datasets, but so far has not been evaluated for image segmentation tasks. In this paper, we tested whether the mixup algorithm can improve the generalization performance of deep segmentation networks for medical image data. We trained a standard U-net architecture to segment the prostate in 100 T2-weighted 3D magnetic resonance images from prostate cancer patients, and compared the results with and without mixup in terms of Dice similarity coefficient and mean surface distance from a reference segmentation made by an experienced radiologist. Our results suggest that mixup offers a statistically significant boost in performance compared to non-mixup training, leading to up to 1.9% increase in Dice and a 10.9% decrease in surface distance. The mixup algorithm may thus offer an important aid for medical image segmentation applications, which are typically limited by severe data scarcity.
摘要研究人员主要通过广泛使用随机翻转、旋转和变形等数据增强技术来解决深度图像处理网络的泛化问题。最近为深度分类网络提出了一种称为mixup的数据增强技术,该技术从输入的凸组合中构建虚拟训练样本。该算法有助于提高各种数据集的分类性能,但迄今为止尚未对图像分割任务进行评估。在本文中,我们测试了混合算法是否可以提高深度分割网络对医学图像数据的泛化性能。我们训练了一个标准的U-net结构来分割来自前列腺癌症患者的100张T2加权3D磁共振图像中的前列腺,并在Dice相似系数和来自经验丰富的放射科医生进行的参考分割的平均表面距离方面比较了有无混淆的结果。我们的研究结果表明,与非混合训练相比,混合训练在统计学上显著提高了成绩,导致骰子增加了1.9%,表面距离减少了10.9%。因此,混合算法可以为医学图像分割应用提供重要的帮助,医学图像分割通常受到严重数据短缺的限制。
{"title":"Mixup (Sample Pairing) Can Improve the Performance of Deep Segmentation Networks","authors":"L. Isaksson, P. Summers, S. Raimondi, S. Gandini, A. Bhalerao, G. Marvaso, G. Petralia, M. Pepa, B. Jereczek-Fossa","doi":"10.2478/jaiscr-2022-0003","DOIUrl":"https://doi.org/10.2478/jaiscr-2022-0003","url":null,"abstract":"Abstract Researchers address the generalization problem of deep image processing networks mainly through extensive use of data augmentation techniques such as random flips, rotations, and deformations. A data augmentation technique called mixup, which constructs virtual training samples from convex combinations of inputs, was recently proposed for deep classification networks. The algorithm contributed to increased performance on classification in a variety of datasets, but so far has not been evaluated for image segmentation tasks. In this paper, we tested whether the mixup algorithm can improve the generalization performance of deep segmentation networks for medical image data. We trained a standard U-net architecture to segment the prostate in 100 T2-weighted 3D magnetic resonance images from prostate cancer patients, and compared the results with and without mixup in terms of Dice similarity coefficient and mean surface distance from a reference segmentation made by an experienced radiologist. Our results suggest that mixup offers a statistically significant boost in performance compared to non-mixup training, leading to up to 1.9% increase in Dice and a 10.9% decrease in surface distance. The mixup algorithm may thus offer an important aid for medical image segmentation applications, which are typically limited by severe data scarcity.","PeriodicalId":48494,"journal":{"name":"Journal of Artificial Intelligence and Soft Computing Research","volume":"12 1","pages":"29 - 39"},"PeriodicalIF":2.8,"publicationDate":"2021-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44857137","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
A Novel Grid-Based Clustering Algorithm 一种新的网格聚类算法
IF 2.8 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2021-10-01 DOI: 10.2478/jaiscr-2021-0019
Artur Starczewski, Magdalena M. Scherer, Wojciech Książek, M. Dębski, Lipo Wang
Abstract Data clustering is an important method used to discover naturally occurring structures in datasets. One of the most popular approaches is the grid-based concept of clustering algorithms. This kind of method is characterized by a fast processing time and it can also discover clusters of arbitrary shapes in datasets. These properties allow these methods to be used in many different applications. Researchers have created many versions of the clustering method using the grid-based approach. However, the key issue is the right choice of the number of grid cells. This paper proposes a novel grid-based algorithm which uses a method for an automatic determining of the number of grid cells. This method is based on the kdist function which computes the distance between each element of a dataset and its kth nearest neighbor. Experimental results have been obtained for several different datasets and they confirm a very good performance of the newly proposed method.
数据聚类是发现数据集中自然结构的一种重要方法。最流行的方法之一是基于网格的聚类算法概念。该方法具有处理速度快,可以发现数据集中任意形状的聚类的特点。这些属性允许在许多不同的应用程序中使用这些方法。研究人员已经使用基于网格的方法创建了许多版本的聚类方法。然而,关键问题是正确选择网格单元的数量。本文提出了一种新的基于网格的算法,该算法使用一种自动确定网格单元数的方法。该方法基于kdist函数,该函数计算数据集的每个元素与其第k个最近邻居之间的距离。在多个不同的数据集上进行了实验,结果证实了该方法的良好性能。
{"title":"A Novel Grid-Based Clustering Algorithm","authors":"Artur Starczewski, Magdalena M. Scherer, Wojciech Książek, M. Dębski, Lipo Wang","doi":"10.2478/jaiscr-2021-0019","DOIUrl":"https://doi.org/10.2478/jaiscr-2021-0019","url":null,"abstract":"Abstract Data clustering is an important method used to discover naturally occurring structures in datasets. One of the most popular approaches is the grid-based concept of clustering algorithms. This kind of method is characterized by a fast processing time and it can also discover clusters of arbitrary shapes in datasets. These properties allow these methods to be used in many different applications. Researchers have created many versions of the clustering method using the grid-based approach. However, the key issue is the right choice of the number of grid cells. This paper proposes a novel grid-based algorithm which uses a method for an automatic determining of the number of grid cells. This method is based on the kdist function which computes the distance between each element of a dataset and its kth nearest neighbor. Experimental results have been obtained for several different datasets and they confirm a very good performance of the newly proposed method.","PeriodicalId":48494,"journal":{"name":"Journal of Artificial Intelligence and Soft Computing Research","volume":"11 1","pages":"319 - 330"},"PeriodicalIF":2.8,"publicationDate":"2021-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42300029","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Decision Making Support System for Managing Advertisers By Ad Fraud Detection 基于广告欺诈检测的广告主管理决策支持系统
IF 2.8 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2021-10-01 DOI: 10.2478/jaiscr-2021-0020
Marcin Gabryel, Magdalena M. Scherer, Ł. Sułkowski, Robertas Damaševičius
Abstract Efficient lead management allows substantially enhancing online channel marketing programs. In the paper, we classify website traffic into human- and bot-origin ones. We use feedforward neural networks with embedding layers. Moreover, we use one-hot encoding for categorical data. The data of mouse clicks come from seven large retail stores and the data of lead classification from three financial institutions. The data are collected by a JavaScript code embedded into HTML pages. The three proposed models achieved relatively high accuracy in detecting artificially generated traffic.
摘要高效的潜在客户管理可以大大增强在线渠道营销计划。在本文中,我们将网站流量分为人类流量和机器人流量。我们使用带有嵌入层的前馈神经网络。此外,我们对分类数据使用一个热编码。鼠标点击数据来自7家大型零售店,线索分类数据来自3家金融机构。数据是通过嵌入HTML页面的JavaScript代码收集的。所提出的三个模型在检测人工生成的流量方面实现了相对较高的精度。
{"title":"Decision Making Support System for Managing Advertisers By Ad Fraud Detection","authors":"Marcin Gabryel, Magdalena M. Scherer, Ł. Sułkowski, Robertas Damaševičius","doi":"10.2478/jaiscr-2021-0020","DOIUrl":"https://doi.org/10.2478/jaiscr-2021-0020","url":null,"abstract":"Abstract Efficient lead management allows substantially enhancing online channel marketing programs. In the paper, we classify website traffic into human- and bot-origin ones. We use feedforward neural networks with embedding layers. Moreover, we use one-hot encoding for categorical data. The data of mouse clicks come from seven large retail stores and the data of lead classification from three financial institutions. The data are collected by a JavaScript code embedded into HTML pages. The three proposed models achieved relatively high accuracy in detecting artificially generated traffic.","PeriodicalId":48494,"journal":{"name":"Journal of Artificial Intelligence and Soft Computing Research","volume":"11 1","pages":"331 - 339"},"PeriodicalIF":2.8,"publicationDate":"2021-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45657960","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A Novel Fast Feedforward Neural Networks Training Algorithm 一种新的快速前馈神经网络训练算法
IF 2.8 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2021-10-01 DOI: 10.2478/jaiscr-2021-0017
J. Bilski, Bartosz Kowalczyk, A. Marjański, M. Gandor, J. Zurada
Abstract In this paper1 a new neural networks training algorithm is presented. The algorithm originates from the Recursive Least Squares (RLS) method commonly used in adaptive filtering. It uses the QR decomposition in conjunction with the Givens rotations for solving a normal equation - resulting from minimization of the loss function. An important parameter in neural networks is training time. Many commonly used algorithms require a big number of iterations in order to achieve a satisfactory outcome while other algorithms are effective only for small neural networks. The proposed solution is characterized by a very short convergence time compared to the well-known backpropagation method and its variants. The paper contains a complete mathematical derivation of the proposed algorithm. There are presented extensive simulation results using various benchmarks including function approximation, classification, encoder, and parity problems. Obtained results show the advantages of the featured algorithm which outperforms commonly used recent state-of-the-art neural networks training algorithms, including the Adam optimizer and the Nesterov’s accelerated gradient.
摘要本文提出了一种新的神经网络训练算法。该算法源于自适应滤波中常用的递归最小二乘法。它使用QR分解和Givens旋转来求解一个正常方程——这是损失函数最小化的结果。神经网络中的一个重要参数是训练时间。许多常用的算法需要大量的迭代才能获得令人满意的结果,而其他算法仅对小型神经网络有效。与众所周知的反向传播方法及其变体相比,所提出的解决方案的特点是收敛时间非常短。本文对所提出的算法进行了完整的数学推导。使用各种基准,包括函数近似、分类、编码器和奇偶校验问题,给出了广泛的模拟结果。获得的结果显示了该特征算法的优势,它优于最近常用的最先进的神经网络训练算法,包括Adam优化器和Nesterov的加速梯度。
{"title":"A Novel Fast Feedforward Neural Networks Training Algorithm","authors":"J. Bilski, Bartosz Kowalczyk, A. Marjański, M. Gandor, J. Zurada","doi":"10.2478/jaiscr-2021-0017","DOIUrl":"https://doi.org/10.2478/jaiscr-2021-0017","url":null,"abstract":"Abstract In this paper1 a new neural networks training algorithm is presented. The algorithm originates from the Recursive Least Squares (RLS) method commonly used in adaptive filtering. It uses the QR decomposition in conjunction with the Givens rotations for solving a normal equation - resulting from minimization of the loss function. An important parameter in neural networks is training time. Many commonly used algorithms require a big number of iterations in order to achieve a satisfactory outcome while other algorithms are effective only for small neural networks. The proposed solution is characterized by a very short convergence time compared to the well-known backpropagation method and its variants. The paper contains a complete mathematical derivation of the proposed algorithm. There are presented extensive simulation results using various benchmarks including function approximation, classification, encoder, and parity problems. Obtained results show the advantages of the featured algorithm which outperforms commonly used recent state-of-the-art neural networks training algorithms, including the Adam optimizer and the Nesterov’s accelerated gradient.","PeriodicalId":48494,"journal":{"name":"Journal of Artificial Intelligence and Soft Computing Research","volume":"11 1","pages":"287 - 306"},"PeriodicalIF":2.8,"publicationDate":"2021-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49585748","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
A New Statistical Reconstruction Method for the Computed Tomography Using an X-Ray Tube with Flying Focal Spot 一种新的x射线管飞焦点ct统计重建方法
IF 2.8 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2021-10-01 DOI: 10.2478/jaiscr-2021-0016
R. Cierniak, P. Pluta, M. Waligóra, Z. Szymanski, K. Grzanek, Filip Pałka, V. Piuri
Abstract This paper presents a new image reconstruction method for spiral cone- beam tomography scanners in which an X-ray tube with a flying focal spot is used. The method is based on principles related to the statistical model-based iterative reconstruction (MBIR) methodology. The proposed approach is a continuous-to-continuous data model approach, and the forward model is formulated as a shift-invariant system. This allows for avoiding a nutating reconstruction-based approach, e.g. the advanced single slice rebinning methodology (ASSR) that is usually applied in computed tomography (CT) scanners with X-ray tubes with a flying focal spot. In turn, the proposed approach allows for significantly accelerating the reconstruction processing and, generally, for greatly simplifying the entire reconstruction procedure. Additionally, it improves the quality of the reconstructed images in comparison to the traditional algorithms, as confirmed by extensive simulations. It is worth noting that the main purpose of introducing statistical reconstruction methods to medical CT scanners is the reduction of the impact of measurement noise on the quality of tomography images and, consequently, the dose reduction of X-ray radiation absorbed by a patient. A series of computer simulations followed by doctor’s assessments have been performed, which indicate how great a reduction of the absorbed dose can be achieved using the reconstruction approach presented here.
摘要本文提出了一种新的螺旋锥束断层扫描仪图像重建方法,该方法采用带飞行焦斑的x射线管。该方法基于基于统计模型的迭代重建(MBIR)方法的相关原理。提出的方法是一种连续到连续的数据模型方法,并将前向模型表述为移位不变系统。这可以避免基于章动重建的方法,例如,通常应用于带有飞行焦斑的x射线管的计算机断层扫描(CT)扫描仪的高级单片重建方法(ASSR)。反过来,所建议的方法可以大大加快重建处理,并且通常大大简化整个重建程序。此外,与传统算法相比,它提高了重建图像的质量,大量的仿真结果证实了这一点。值得注意的是,将统计重建方法引入医用CT扫描仪的主要目的是减少测量噪声对断层扫描图像质量的影响,从而减少患者吸收的x射线辐射剂量。在医生的评估之后进行了一系列的计算机模拟,这表明使用这里提出的重建方法可以实现吸收剂量的减少。
{"title":"A New Statistical Reconstruction Method for the Computed Tomography Using an X-Ray Tube with Flying Focal Spot","authors":"R. Cierniak, P. Pluta, M. Waligóra, Z. Szymanski, K. Grzanek, Filip Pałka, V. Piuri","doi":"10.2478/jaiscr-2021-0016","DOIUrl":"https://doi.org/10.2478/jaiscr-2021-0016","url":null,"abstract":"Abstract This paper presents a new image reconstruction method for spiral cone- beam tomography scanners in which an X-ray tube with a flying focal spot is used. The method is based on principles related to the statistical model-based iterative reconstruction (MBIR) methodology. The proposed approach is a continuous-to-continuous data model approach, and the forward model is formulated as a shift-invariant system. This allows for avoiding a nutating reconstruction-based approach, e.g. the advanced single slice rebinning methodology (ASSR) that is usually applied in computed tomography (CT) scanners with X-ray tubes with a flying focal spot. In turn, the proposed approach allows for significantly accelerating the reconstruction processing and, generally, for greatly simplifying the entire reconstruction procedure. Additionally, it improves the quality of the reconstructed images in comparison to the traditional algorithms, as confirmed by extensive simulations. It is worth noting that the main purpose of introducing statistical reconstruction methods to medical CT scanners is the reduction of the impact of measurement noise on the quality of tomography images and, consequently, the dose reduction of X-ray radiation absorbed by a patient. A series of computer simulations followed by doctor’s assessments have been performed, which indicate how great a reduction of the absorbed dose can be achieved using the reconstruction approach presented here.","PeriodicalId":48494,"journal":{"name":"Journal of Artificial Intelligence and Soft Computing Research","volume":"11 1","pages":"271 - 286"},"PeriodicalIF":2.8,"publicationDate":"2021-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43331395","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
Journal of Artificial Intelligence and Soft Computing Research
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1