首页 > 最新文献

Radio Electronics Computer Science Control最新文献

英文 中文
NONLINEAR REGRESSION MODELS FOR ESTIMATING THE DURATION OF SOFTWARE DEVELOPMENT IN JAVA FOR PC BASED ON THE 2021 ISBSG DATA 基于2021 isbsg数据估算PC Java软件开发周期的非线性回归模型
IF 0.5 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2022-10-17 DOI: 10.15588/1607-3274-2022-3-14
S. Prykhodko, A. Pukhalevych, K. Prykhodko, L. Makarova
Context. The problem of estimating the duration of software development in Java for personal computers (PC) is important because, first, failed duration estimating is often the main contributor to failed software projects, second, Java is a popular language, and, third, a personal computer is a widespread multi-purpose computer. The object of the study is the process of estimating the duration of software development in Java for PC. The subject of the study is the nonlinear regression models to estimate the duration of software development in Java for PC. Objective. The goal of the work is to build nonlinear regression models for estimating the duration of software development in Java for PC based on the normalizing transformations and deleting outliers in data to increase the confidence of the estimation in comparison to the ISBSG model for the PC platform. Method. The models, confidence, and prediction intervals of nonlinear regressions to estimate the duration of software development in Java for PC are constructed based on the normalizing transformations for non-Gaussian data with the help of appropriate techniques. The techniques to build the models, confidence, and prediction intervals of nonlinear regressions are based on normalizing transformations. Also, we apply outlier removal for model construction. In general, the above leads to a reduction of the mean magnitude of relative error, the widths of the confidence, and prediction intervals in comparison to nonlinear models constructed without outlier removal application in the model construction process. Results. A comparison of the model based on the decimal logarithm transformation with the nonlinear regression models based on the Johnson (for the SB family) and Box-Cox transformations as both univariate and bivariate ones has been performed. Conclusions. The nonlinear regression model to estimate the duration of software development in Java for PC is constructed based on the decimal logarithm transformation. This model, in comparison with other nonlinear regression models, has smaller widths of the confidence and prediction intervals for effort values that are bigger than 900 person-hours. The prospects for further research may include the application of bivariate normalizing transformations and data sets to construct the nonlinear regression models for estimating the duration of software development in other languages for PC and other platforms, for example, mainframe.
上下文。估计用于个人计算机(PC)的Java软件开发持续时间的问题很重要,因为首先,失败的持续时间估计通常是导致软件项目失败的主要原因,其次,Java是一种流行的语言,第三,个人计算机是一种广泛使用的多用途计算机。本研究的对象是估算PC Java软件开发周期的过程。本研究的主题是用非线性回归模型来估计PC版Java软件开发的持续时间。目标。这项工作的目标是建立非线性回归模型,用于估计PC平台Java软件开发的持续时间,该模型基于规范化转换和删除数据中的异常值,以增加与PC平台ISBSG模型相比估计的置信度。方法。基于非高斯数据的归一化转换,在适当的技术帮助下,构建了用于估计PC Java软件开发持续时间的非线性回归模型、置信度和预测区间。建立非线性回归模型、置信度和预测区间的技术是基于归一化变换的。此外,我们还将异常值去除用于模型构建。总的来说,与在模型构建过程中没有应用离群值去除的非线性模型相比,上述方法导致相对误差的平均幅度、置信宽度和预测区间减小。结果。将基于十进制对数变换的模型与基于Johnson(适用于SB家族)和Box-Cox变换的非线性回归模型作为单变量和双变量模型进行了比较。结论。基于十进制对数变换,建立了用于估算PC Java软件开发周期的非线性回归模型。与其他非线性回归模型相比,该模型对于大于900人小时的工作量值具有较小的置信宽度和预测区间。进一步研究的前景可能包括应用二元归一化转换和数据集来构建非线性回归模型,以估计用于PC和其他平台(例如大型机)的其他语言软件开发的持续时间。
{"title":"NONLINEAR REGRESSION MODELS FOR ESTIMATING THE DURATION OF SOFTWARE DEVELOPMENT IN JAVA FOR PC BASED ON THE 2021 ISBSG DATA","authors":"S. Prykhodko, A. Pukhalevych, K. Prykhodko, L. Makarova","doi":"10.15588/1607-3274-2022-3-14","DOIUrl":"https://doi.org/10.15588/1607-3274-2022-3-14","url":null,"abstract":"Context. The problem of estimating the duration of software development in Java for personal computers (PC) is important because, first, failed duration estimating is often the main contributor to failed software projects, second, Java is a popular language, and, third, a personal computer is a widespread multi-purpose computer. The object of the study is the process of estimating the duration of software development in Java for PC. The subject of the study is the nonlinear regression models to estimate the duration of software development in Java for PC. \u0000Objective. The goal of the work is to build nonlinear regression models for estimating the duration of software development in Java for PC based on the normalizing transformations and deleting outliers in data to increase the confidence of the estimation in comparison to the ISBSG model for the PC platform. \u0000Method. The models, confidence, and prediction intervals of nonlinear regressions to estimate the duration of software development in Java for PC are constructed based on the normalizing transformations for non-Gaussian data with the help of appropriate techniques. The techniques to build the models, confidence, and prediction intervals of nonlinear regressions are based on normalizing transformations. Also, we apply outlier removal for model construction. In general, the above leads to a reduction of the mean magnitude of relative error, the widths of the confidence, and prediction intervals in comparison to nonlinear models constructed without outlier removal application in the model construction process. \u0000Results. A comparison of the model based on the decimal logarithm transformation with the nonlinear regression models based on the Johnson (for the SB family) and Box-Cox transformations as both univariate and bivariate ones has been performed. \u0000Conclusions. The nonlinear regression model to estimate the duration of software development in Java for PC is constructed based on the decimal logarithm transformation. This model, in comparison with other nonlinear regression models, has smaller widths of the confidence and prediction intervals for effort values that are bigger than 900 person-hours. The prospects for further research may include the application of bivariate normalizing transformations and data sets to construct the nonlinear regression models for estimating the duration of software development in other languages for PC and other platforms, for example, mainframe.","PeriodicalId":43783,"journal":{"name":"Radio Electronics Computer Science Control","volume":"25 1","pages":""},"PeriodicalIF":0.5,"publicationDate":"2022-10-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85107064","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MULTILINGUAL TEXT CLASSIFIER USING PRE-TRAINED UNIVERSAL SENTENCE ENCODER MODEL 使用预训练通用句子编码器模型的多语言文本分类器
IF 0.5 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2022-10-16 DOI: 10.15588/1607-3274-2022-3-10
O. V. Orlovskiy, K. Sohrab, S. Ostapov, K. P. Hazdyuk, L. Shumylyak
Context. Online platforms and environments continue to generate ever-increasing content. The task of automating the moderation of user-generated content continues to be relevant. Of particular note are cases in which, for one reason or another, there is a very small amount of data to teach the classifier. To achieve results under such conditions, it is important to involve the classifier pre-trained models, which were trained on a large amount of data from a wide range. This paper deals with the use of the pre-trained multilingual Universal Sentence Encoder (USE) model as a component of the developed classifier and the affect of hyperparameters on the classification accuracy when learning on a small data amount (~ 0.05% of the dataset). Objective. The goal of this paper is the investigation of the pre-trained multilingual model and optimal hyperparameters influence for learning the text data classifier on the classification result. Method. To solve this problem, a relatively new approach to few-shot learning has recently been used – learning with a relatively small number of examples. Since text data is still the dominant way of transmitting information, the study of the possibilities of constructing a classifier of text data when learning from a small number of examples (~ 0.002–0.05% of the data set) is an actual problem. Results. It is shown that even with a small number of examples for learning (36 per class) due to the use of USE and optimal configuration in learning can achieve high accuracy of classification on English and Russian data, which is extremely important when it is impossible to collect your own large data set. The influence of the approach using USE and a set of different configurations of hyperparameters on the result of the text data classifier on the example of English and Russian data sets is evaluated. Conclusions. During the experiments, a significant degree of relevance of the correct selection of hyperparameters is shown. In particular, this paper considered the batch size, optimizer, number of learning epochs and the percentage of data from the set taken to train the classifier. In the process of experimentation, the optimal configuration of hyperparameters was selected, according to which 86.46% accuracy of classification on the Russian-language data set and 91.13% on the English-language data, respectively, can be achieved in ten seconds of training (training time can be significantly affected by technical means used).
上下文。网络平台和环境不断产生越来越多的内容。自动审核用户生成内容的任务仍然具有相关性。特别值得注意的是,由于这样或那样的原因,只有非常少的数据可以教分类器。为了在这种情况下获得结果,涉及分类器预训练模型是很重要的,这些模型是在广泛的大量数据上训练的。本文讨论了使用预训练的多语言通用句子编码器(use)模型作为开发的分类器的组成部分,以及在小数据量(约0.05%的数据集)上学习时超参数对分类精度的影响。目标。本文的目的是研究预训练的多语言模型和学习文本数据分类器的最优超参数对分类结果的影响。方法。为了解决这个问题,最近使用了一种相对较新的少采样学习方法——使用相对较少的样本进行学习。由于文本数据仍然是传递信息的主要方式,研究在从少量示例(约0.002-0.05%的数据集)中学习时构建文本数据分类器的可能性是一个实际问题。结果。研究表明,由于使用use和学习中的最佳配置,即使使用少量的示例进行学习(每类36个),也可以实现英语和俄语数据的高精度分类,这在不可能收集自己的大数据集时非常重要。以英语和俄语数据集为例,评估了使用USE和一组不同超参数配置的方法对文本数据分类器结果的影响。结论。在实验中,超参数的正确选择具有显著的相关性。特别地,本文考虑了批处理大小、优化器、学习周期数和用于训练分类器的数据集的百分比。在实验过程中,选择了最优的超参数配置,在10秒的训练时间内,俄语数据集的分类准确率达到86.46%,英语数据集的分类准确率达到91.13%(使用的技术手段对训练时间有显著影响)。
{"title":"MULTILINGUAL TEXT CLASSIFIER USING PRE-TRAINED UNIVERSAL SENTENCE ENCODER MODEL","authors":"O. V. Orlovskiy, K. Sohrab, S. Ostapov, K. P. Hazdyuk, L. Shumylyak","doi":"10.15588/1607-3274-2022-3-10","DOIUrl":"https://doi.org/10.15588/1607-3274-2022-3-10","url":null,"abstract":"Context. Online platforms and environments continue to generate ever-increasing content. The task of automating the moderation of user-generated content continues to be relevant. Of particular note are cases in which, for one reason or another, there is a very small amount of data to teach the classifier. To achieve results under such conditions, it is important to involve the classifier pre-trained models, which were trained on a large amount of data from a wide range. This paper deals with the use of the pre-trained multilingual Universal Sentence Encoder (USE) model as a component of the developed classifier and the affect of hyperparameters on the classification accuracy when learning on a small data amount (~ 0.05% of the dataset). \u0000Objective. The goal of this paper is the investigation of the pre-trained multilingual model and optimal hyperparameters influence for learning the text data classifier on the classification result. \u0000Method. To solve this problem, a relatively new approach to few-shot learning has recently been used – learning with a relatively small number of examples. Since text data is still the dominant way of transmitting information, the study of the possibilities of constructing a classifier of text data when learning from a small number of examples (~ 0.002–0.05% of the data set) is an actual problem. \u0000Results. It is shown that even with a small number of examples for learning (36 per class) due to the use of USE and optimal configuration in learning can achieve high accuracy of classification on English and Russian data, which is extremely important when it is impossible to collect your own large data set. The influence of the approach using USE and a set of different configurations of hyperparameters on the result of the text data classifier on the example of English and Russian data sets is evaluated. \u0000Conclusions. During the experiments, a significant degree of relevance of the correct selection of hyperparameters is shown. In particular, this paper considered the batch size, optimizer, number of learning epochs and the percentage of data from the set taken to train the classifier. In the process of experimentation, the optimal configuration of hyperparameters was selected, according to which 86.46% accuracy of classification on the Russian-language data set and 91.13% on the English-language data, respectively, can be achieved in ten seconds of training (training time can be significantly affected by technical means used).","PeriodicalId":43783,"journal":{"name":"Radio Electronics Computer Science Control","volume":"51 1","pages":""},"PeriodicalIF":0.5,"publicationDate":"2022-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85813977","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MATHEMATICAL MODEL FOR DECISION MAKING SYSTEM BASED ON THREE-SEGMENTED LINEAR REGRESSION 基于三段线性回归的决策系统数学模型
IF 0.5 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2022-10-16 DOI: 10.15588/1607-3274-2022-3-4
V. Kuzmin, R. Khrashchevskyi, M. Kulik, O. Ivanets, M. Zaliskyi, Yu. V. Petrova
Context. The problem of approximation of empirical data in the decision-making system in safety management.. The object of the study was to verify the adequate coefficients of the mathematical model for data approximation using information technology. Objective. The goal of the work is the creation adequate math-ematical model using information technology on the bases analyze different approaches for approximating empirical data an that can be used to predict the current state of the operator in the flight safety system.. Method. A comparative analysis of the description of the transformation of information indicators with a non-standard structure. The following models of transformation of information indicators with similar visual representation are selected for comparison: parabolas of the second and third order, single regression and regression with jumps. It is proposed to use new approaches for approximation, based on the use of the criterion proposed by Kuzmin and the Heaviside function. The adequacy of the approximation was checked using these criteria, which allowed to choose an adequate mathematical model to describe the transformation of information indicators. The stages of obtaining a mathematical model were as follows: determining the minimum sum of squares of deviations for all information indicators simultaneously; use of the Heaviside function; optimization of the abscissa axis in certain areas; use of the linearity test. The obtained mathematical model adequately describes the process of transformation of information indicators, which will allow the process of forecasting changes in medical and biological indicators of operators in the performance of professional duties in aviation, as one of the methods of determining the human factor in a proactive approach in flight safety. Results. The results of the study can be used during the construction of mathematical models to describe empirical data of this kind. Conclusions. Experimental studies have suggested recommending the use of three-segment linear regression with jumps as an adequate mathematical model that can be used to formalize the description of empirical data with non-standard structure and can be used in practice to build models for predicting operator dysfunction as one of the causes of adverse events in aviation. Prospects for further research may be the creation of a multiparameter mathematical model that will predict the violation of the functional state of the operator by informative parameters, as well as experimental study of proposed mathematical approaches for a wide range of practical problems of different nature and dimension.
上下文。安全管理决策系统中经验数据的逼近问题研究的目的是验证利用信息技术进行数据逼近的数学模型的适当系数。目标。工作的目标是利用信息技术在分析近似经验数据的不同方法的基础上创建适当的数学模型,该模型可用于预测飞行安全系统中操作员的当前状态。方法。非标准结构下信息指标转换描述的比较分析选择具有相似视觉表征的信息指标转换模型进行比较:二阶抛物线和三阶抛物线、单次回归和跳跃回归。在使用Kuzmin提出的准则和Heaviside函数的基础上,提出了新的逼近方法。使用这些标准检查了近似的充分性,从而可以选择适当的数学模型来描述信息指标的转换。获得数学模型的步骤如下:同时确定所有信息指标偏差的最小平方和;使用Heaviside功能;部分区域横坐标的优化;使用线性测试。所获得的数学模型充分描述了信息指标的转换过程,这将使预测操作人员在履行航空专业职责时的医疗和生物指标变化的过程成为确定飞行安全主动办法中的人为因素的方法之一。结果。研究结果可用于构建描述此类实证数据的数学模型。结论。实验研究建议使用带有跳跃的三段线性回归作为一种适当的数学模型,可以用来形式化描述具有非标准结构的经验数据,并且可以在实践中用于构建预测操作员功能障碍作为航空不良事件原因之一的模型。进一步研究的前景可能是建立一个多参数数学模型,该模型将通过信息参数预测算子的功能状态的破坏,以及对各种不同性质和维度的实际问题提出的数学方法的实验研究。
{"title":"MATHEMATICAL MODEL FOR DECISION MAKING SYSTEM BASED ON THREE-SEGMENTED LINEAR REGRESSION","authors":"V. Kuzmin, R. Khrashchevskyi, M. Kulik, O. Ivanets, M. Zaliskyi, Yu. V. Petrova","doi":"10.15588/1607-3274-2022-3-4","DOIUrl":"https://doi.org/10.15588/1607-3274-2022-3-4","url":null,"abstract":"Context. The problem of approximation of empirical data in the decision-making system in safety management.. The object of the study was to verify the adequate coefficients of the mathematical model for data approximation using information technology. \u0000Objective. The goal of the work is the creation adequate math-ematical model using information technology on the bases analyze different approaches for approximating empirical data an that can be used to predict the current state of the operator in the flight safety system.. \u0000Method. A comparative analysis of the description of the transformation of information indicators with a non-standard structure. The following models of transformation of information indicators with similar visual representation are selected for comparison: parabolas of the second and third order, single regression and regression with jumps. It is proposed to use new approaches for approximation, based on the use of the criterion proposed by Kuzmin and the Heaviside function. The adequacy of the approximation was checked using these criteria, which allowed to choose an adequate mathematical model to describe the transformation of information indicators. The stages of obtaining a mathematical model were as follows: determining the minimum sum of squares of deviations for all information indicators simultaneously; use of the Heaviside function; optimization of the abscissa axis in certain areas; use of the linearity test. The obtained mathematical model adequately describes the process of transformation of information indicators, which will allow the process of forecasting changes in medical and biological indicators of operators in the performance of professional duties in aviation, as one of the methods of determining the human factor in a proactive approach in flight safety. \u0000Results. The results of the study can be used during the construction of mathematical models to describe empirical data of this kind. \u0000Conclusions. Experimental studies have suggested recommending the use of three-segment linear regression with jumps as an adequate mathematical model that can be used to formalize the description of empirical data with non-standard structure and can be used in practice to build models for predicting operator dysfunction as one of the causes of adverse events in aviation. \u0000Prospects for further research may be the creation of a multiparameter mathematical model that will predict the violation of the functional state of the operator by informative parameters, as well as experimental study of proposed mathematical approaches for a wide range of practical problems of different nature and dimension.","PeriodicalId":43783,"journal":{"name":"Radio Electronics Computer Science Control","volume":"14 1","pages":""},"PeriodicalIF":0.5,"publicationDate":"2022-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81939470","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
CREDIBILISTIC FUZZY CLUSTERING BASED ON ANALYSIS OF DATA DISTRIBUTION DENSITY AND THEIR PEAKS 基于数据分布密度及其峰值分析的可信模糊聚类
IF 0.5 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2022-10-16 DOI: 10.15588/1607-3274-2022-3-6
Y. Bodyanskiy, I. Pliss, A. Shafronenko, O. Kalynychenko
Context. The task of clustering – classification without a teacher of data arrays occupies a rather important place in Data Mining. To solve this problem, many approaches have been proposed at the moment, differing from each other in a priori assumptions in the studied and analyzed arrays, in the mathematical apparatus that is the basis of certain methods. The solution of clustering problems is complicated by the large dimension of the vectors of the analyzed observations, their distortion of various types. Objective. The purpose of the work is to introduce a fuzzy clustering procedure that combines the advantages of methods based on the analysis of data distribution densities and their peaks, which are characterized by high speed and can work effectively in conditions of classes that overlapping. Method. The method of fuzzy clustering of data arrays, based on the ideas of analyzing the distribution densities of these data, their peaks, and a confidence fuzzy approach has been introduced. The advantage of the proposed approach is to reduce the time for solving optimization problems related to finding attractors of density functions, since the number of calls to the optimization block is determined not by the volume of the analyzed array, but by the number of density peaks of the same array. Results. The method is quite simple in numerical implementation and is not critical to the choice of the optimization procedure. The experimental results confirm the effectiveness of the proposed approach in clustering problems under the condition of cluster intersection and allow us to recommend the proposed method for practical use in solving problems of automatic clustering of large data volumes. Conclusions. The method is quite simple in numerical implementation and is not critical to the choice of the optimization procedure. The advantage of the proposed approach is to reduce the time for solving optimization problems related to finding attractors of density functions, since the number of calls to the optimization block is determined not by the volume of the analyzed array, but by the number of density peaks of the same array. The method is quite simple in numerical implementation and is not critical to the choice of the optimization procedure. The experimental results confirm the effectiveness of the proposed approach in clustering problems under conditions of overlapping clusters.
上下文。在数据挖掘中,没有数据数组老师的聚类分类任务占有相当重要的地位。为了解决这个问题,目前已经提出了许多方法,在研究和分析数组的先验假设中,在作为某些方法基础的数学装置中,彼此不同。聚类问题的求解由于所分析的观测值的向量维数较大,且存在各种类型的畸变而变得复杂。目标。本文的目的是引入一种模糊聚类方法,该方法结合了基于数据分布密度及其峰值分析的方法的优点,该方法具有速度快的特点,可以有效地在类重叠的情况下工作。方法。在分析数据的分布密度、峰值和置信度模糊方法的基础上,提出了数据阵列的模糊聚类方法。所提出的方法的优点是减少了解决与寻找密度函数吸引子相关的优化问题的时间,因为对优化块的调用次数不是由分析数组的体积决定的,而是由同一数组的密度峰的数量决定的。结果。该方法的数值实现非常简单,对优化过程的选择也不重要。实验结果证实了该方法在聚类相交条件下的聚类问题中的有效性,并为实际应用中解决大数据量自动聚类问题提供了参考。结论。该方法的数值实现非常简单,对优化过程的选择也不重要。所提出的方法的优点是减少了解决与寻找密度函数吸引子相关的优化问题的时间,因为对优化块的调用次数不是由分析数组的体积决定的,而是由同一数组的密度峰的数量决定的。该方法的数值实现非常简单,对优化过程的选择也不重要。实验结果证实了该方法在重叠聚类条件下处理聚类问题的有效性。
{"title":"CREDIBILISTIC FUZZY CLUSTERING BASED ON ANALYSIS OF DATA DISTRIBUTION DENSITY AND THEIR PEAKS","authors":"Y. Bodyanskiy, I. Pliss, A. Shafronenko, O. Kalynychenko","doi":"10.15588/1607-3274-2022-3-6","DOIUrl":"https://doi.org/10.15588/1607-3274-2022-3-6","url":null,"abstract":"Context. The task of clustering – classification without a teacher of data arrays occupies a rather important place in Data Mining. To solve this problem, many approaches have been proposed at the moment, differing from each other in a priori assumptions in the studied and analyzed arrays, in the mathematical apparatus that is the basis of certain methods. The solution of clustering problems is complicated by the large dimension of the vectors of the analyzed observations, their distortion of various types. \u0000Objective. The purpose of the work is to introduce a fuzzy clustering procedure that combines the advantages of methods based on the analysis of data distribution densities and their peaks, which are characterized by high speed and can work effectively in conditions of classes that overlapping. \u0000Method. The method of fuzzy clustering of data arrays, based on the ideas of analyzing the distribution densities of these data, their peaks, and a confidence fuzzy approach has been introduced. The advantage of the proposed approach is to reduce the time for solving optimization problems related to finding attractors of density functions, since the number of calls to the optimization block is determined not by the volume of the analyzed array, but by the number of density peaks of the same array. \u0000Results. The method is quite simple in numerical implementation and is not critical to the choice of the optimization procedure. The experimental results confirm the effectiveness of the proposed approach in clustering problems under the condition of cluster intersection and allow us to recommend the proposed method for practical use in solving problems of automatic clustering of large data volumes. \u0000Conclusions. The method is quite simple in numerical implementation and is not critical to the choice of the optimization procedure. The advantage of the proposed approach is to reduce the time for solving optimization problems related to finding attractors of density functions, since the number of calls to the optimization block is determined not by the volume of the analyzed array, but by the number of density peaks of the same array. The method is quite simple in numerical implementation and is not critical to the choice of the optimization procedure. The experimental results confirm the effectiveness of the proposed approach in clustering problems under conditions of overlapping clusters.","PeriodicalId":43783,"journal":{"name":"Radio Electronics Computer Science Control","volume":"165 1","pages":""},"PeriodicalIF":0.5,"publicationDate":"2022-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74881669","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
OPTIMIZATION OF SWARM ROBOTICS ALGORITHMS 群机器人算法的优化
IF 0.5 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2022-10-16 DOI: 10.15588/1607-3274-2022-3-7
Tetiana A. Vakaliuk, R. Kukharchuk, O. Zaika, A. V. Riabko
Context. Among the variety of tasks solved by robotics, one can single out a number of those for the solution of which small dimensions of work are desirable and sometimes necessary. To solve such problems, micro-robots with small dimensions are needed, the mass of which allows them to move freely in tight passages, in difficult weather conditions, and remain unnoticed. At the same time, the small dimensions of the microrobot also impose some indirect restrictions; therefore, it is better to use groups of microrobots for the solution of these problems. The efficiency of using groups of microrobots depends on the chosen control strategy and stochastic search algorithms for optimizing the control of a group (swarm) of microrobots. Objective. The purpose of this work is to consider a group of swarm algorithms (methods) belonging to the class of metaheuristics. The group of these algorithms includes, in particular, the ant colony algorithm, the possibilities of which were investigated to solve the traveling salesman problem, which often arises when developing an algorithm for the behavior of a group of microrobots. Method. At the first stage of the study, the main groups of parameters were identified that determine the flow and characterize the state at any time of the ant colony algorithm: input, control, disturbance parameters, output parameters. After identifying the main groups of parameters, an algorithm was developed, the advantage of which lies in scalability, as well as guaranteed convergence, which makes it possible to obtain an optimal solution regardless of the dimension of the graph. At the second stage, an algorithm was developed, the code of which was implemented in the Matlab language. Computer experiments were carried out to determine the influence of input, control, output, and disturbance parameters on the convergence of the algorithm. Attention was paid to the main groups of indicators that determine the direction of the method and characterize the state of the swarm of microrobots at a given time. In the computational experiment, the number of ants placed in the nodes of the network, the amount of pheromone, the number of graph nodes were varied, the number of iterations to find the shortest path, and the execution time of the method were determined. The final test of modeling and performance of the method was carried out. Results. Research has been carried out on the application of the ant algorithm for solving the traveling salesman problem for test graphs with a random arrangement of vertices; for a constant number of vertices and a change in the number of ants, for a constant number of vertices at different values of the coefficient Q; to solve the traveling salesman problem for a constant number of vertices at different values of the pheromone evaporation coefficient p; for a different number of graph vertices. The results showed that ant methods find good traveling salesman routes much faster than clear-cut combinatorial
上下文。在机器人技术解决的各种任务中,人们可以挑出一些需要小尺寸工作的任务,有时甚至是必要的。为了解决这些问题,需要小尺寸的微型机器人,它们的质量可以让它们在狭窄的通道中自由移动,在恶劣的天气条件下,不被人注意。同时,微型机器人的小尺寸也对其施加了一些间接的限制;因此,最好使用微型机器人群来解决这些问题。使用微机器人群的效率取决于所选择的控制策略和优化微机器人群控制的随机搜索算法。目标。本工作的目的是考虑一组属于元启发式类的群算法(方法)。这些算法特别包括蚁群算法,研究了其解决旅行推销员问题的可能性,这在为一组微型机器人的行为开发算法时经常出现。方法。在研究的第一阶段,确定了蚁群算法中决定流量和表征任意时刻状态的主要参数组:输入参数、控制参数、干扰参数、输出参数。在确定了主要参数组后,开发了一种算法,该算法的优点在于可扩展性和保证收敛性,使得无论图的维数如何,都可以获得最优解。第二阶段,开发了一种算法,并用Matlab语言实现了算法的代码。通过计算机实验确定了输入、控制、输出和干扰参数对算法收敛性的影响。注意了确定方法方向和表征微机器人群在给定时间的状态的主要指标组。在计算实验中,确定了网络节点中蚂蚁的数量、信息素的数量、图节点的数量、寻找最短路径的迭代次数以及该方法的执行时间。最后对该方法进行了建模和性能测试。结果。研究了蚁群算法在随机点排列测试图的旅行商问题中的应用;对于一定数量的顶点和蚂蚁数量的变化,对于不同系数Q值的一定数量的顶点;求解在不同费洛蒙蒸发系数p值下的恒定顶点数下的旅行推销员问题;对于不同数量的图顶点。结果表明,蚁群方法比清晰组合优化方法更快地找到最佳的旅行商路线。通过不同顶点数和迭代次数的测试网络实例,建立了搜索时间和找到的最优路径与控制参数值的依赖关系。结论。这些研究是为了对蚁群算法的应用提出建议,以控制一组(群)微型机器人。
{"title":"OPTIMIZATION OF SWARM ROBOTICS ALGORITHMS","authors":"Tetiana A. Vakaliuk, R. Kukharchuk, O. Zaika, A. V. Riabko","doi":"10.15588/1607-3274-2022-3-7","DOIUrl":"https://doi.org/10.15588/1607-3274-2022-3-7","url":null,"abstract":"Context. Among the variety of tasks solved by robotics, one can single out a number of those for the solution of which small dimensions of work are desirable and sometimes necessary. To solve such problems, micro-robots with small dimensions are needed, the mass of which allows them to move freely in tight passages, in difficult weather conditions, and remain unnoticed. At the same time, the small dimensions of the microrobot also impose some indirect restrictions; therefore, it is better to use groups of microrobots for the solution of these problems. The efficiency of using groups of microrobots depends on the chosen control strategy and stochastic search algorithms for optimizing the control of a group (swarm) of microrobots. \u0000Objective. The purpose of this work is to consider a group of swarm algorithms (methods) belonging to the class of metaheuristics. The group of these algorithms includes, in particular, the ant colony algorithm, the possibilities of which were investigated to solve the traveling salesman problem, which often arises when developing an algorithm for the behavior of a group of microrobots. \u0000Method. At the first stage of the study, the main groups of parameters were identified that determine the flow and characterize the state at any time of the ant colony algorithm: input, control, disturbance parameters, output parameters. After identifying the main groups of parameters, an algorithm was developed, the advantage of which lies in scalability, as well as guaranteed convergence, which makes it possible to obtain an optimal solution regardless of the dimension of the graph. At the second stage, an algorithm was developed, the code of which was implemented in the Matlab language. Computer experiments were carried out to determine the influence of input, control, output, and disturbance parameters on the convergence of the algorithm. Attention was paid to the main groups of indicators that determine the direction of the method and characterize the state of the swarm of microrobots at a given time. In the computational experiment, the number of ants placed in the nodes of the network, the amount of pheromone, the number of graph nodes were varied, the number of iterations to find the shortest path, and the execution time of the method were determined. The final test of modeling and performance of the method was carried out. \u0000Results. Research has been carried out on the application of the ant algorithm for solving the traveling salesman problem for test graphs with a random arrangement of vertices; for a constant number of vertices and a change in the number of ants, for a constant number of vertices at different values of the coefficient Q; to solve the traveling salesman problem for a constant number of vertices at different values of the pheromone evaporation coefficient p; for a different number of graph vertices. The results showed that ant methods find good traveling salesman routes much faster than clear-cut combinatorial ","PeriodicalId":43783,"journal":{"name":"Radio Electronics Computer Science Control","volume":"46 1","pages":""},"PeriodicalIF":0.5,"publicationDate":"2022-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89948780","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IMAGE CLASSIFIER RESILIENT TO ADVERSARIAL ATTACKS, FAULT INJECTIONS AND CONCEPT DRIFT – MODEL ARCHITECTURE AND TRAINING ALGORITHM 对抗攻击、故障注入和概念漂移的图像分类器——模型架构和训练算法
IF 0.5 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2022-10-16 DOI: 10.15588/1607-3274-2022-3-9
V. Moskalenko, A. Moskalenko, A. Korobov, M. O. Zaretsky
Context. The problem of image classification algorithms vulnerability to destructive perturbations has not yet been definitively resolved and is quite relevant for safety-critical applications. Therefore, object of research is the process of training and inference for image classifier that functioning under influences of destructive perturbations. The subjects of the research are model architecture and training algorithm of image classifier that provide resilience to adversarial attacks, fault injection attacks and concept drift. Objective. Stated research goal is to develop effective model architecture and training algorithm that provide resilience to adversarial attacks, fault injections and concept drift. Method. New training algorithm which combines self-knowledge distillation, information measure maximization, class distribution compactness and interclass gap maximization, data compression based on discretization of feature representation and semi-supervised learning based on consistency regularization is proposed. Results. The model architecture and training algorithm of image classifier were developed. The obtained classifier was tested on the Cifar10 dataset to evaluate its resilience over an interval of 200 mini-batches with a training and test size of mini-batch equals to 128 examples for such perturbations: adversarial black-box L∞-attacks with perturbation levels equal to 1, 3, 5 and 10; inversion of one randomly selected bit in a tensor for 10%, 30%, 50% and 60% randomly selected tensors; addition of one new class; real concept drift between a pair of classes. The effect of the feature space dimensionality on the value of the information criterion of the model performance without perturbations and the value of the integral metric of resilience during the exposure to perturbations is considered. Conclusions. The proposed model architecture and learning algorithm provide absorption of part of the disturbing influence, graceful degradation due to hierarchical classes and adaptive computation, and fast adaptation on a limited amount of labeled data. It is shown that adaptive computation saves up to 40% of resources due to early decision-making in the lower sections of the model, but perturbing influence leads to slowing down, which can be considered as graceful degradation. A multi-section structure trained using knowledge self-distillation principles has been shown to provide more than 5% improvement in the value of the integral mectric of resilience compared to an architecture where the decision is made on the last layer of the model. It is observed that the dimensionality of the feature space noticeably affects the resilience to adversarial attacks and can be chosen as a tradeoff between resilience to perturbations and efficiency without perturbations.
上下文。图像分类算法易受破坏性扰动的问题尚未得到明确解决,并且与安全关键应用非常相关。因此,研究对象是在破坏性扰动影响下运行的图像分类器的训练和推理过程。研究的主题是图像分类器的模型架构和训练算法,以提供对抗攻击、故障注入攻击和概念漂移的弹性。目标。声明的研究目标是开发有效的模型架构和训练算法,以提供对抗性攻击,故障注入和概念漂移的弹性。方法。提出了一种结合自知识升华、信息测度最大化、类分布紧密性和类间间隙最大化、基于特征表示离散化的数据压缩和基于一致性正则化的半监督学习的训练算法。结果。提出了图像分类器的模型体系结构和训练算法。得到的分类器在Cifar10数据集上进行测试,以评估其在200个小批次的间隔内的弹性,小批次的训练和测试大小等于128个例子,用于以下扰动:扰动级别为1,3,5和10的对抗性黑盒L∞攻击;在张量中对10%、30%、50%和60%随机选择的张量进行一个随机选择位的反演;增加一个新类别;真正的概念在两个类之间漂移。考虑了特征空间维数对无扰动时模型性能信息准则值和受扰动时弹性积分度量值的影响。结论。所提出的模型结构和学习算法可以吸收部分干扰影响,由于分层类和自适应计算而实现优雅的退化,并且可以对有限数量的标记数据进行快速自适应。结果表明,自适应计算由于模型下部的早期决策而节省了高达40%的资源,但扰动影响导致了减速,可以认为是一种优雅的退化。与在模型的最后一层做出决策的体系结构相比,使用知识自蒸馏原则训练的多部分结构已被证明提供了超过5%的弹性积分度量值的改进。观察到特征空间的维数显著影响对抗性攻击的弹性,并且可以作为对扰动的弹性和无扰动的效率之间的权衡。
{"title":"IMAGE CLASSIFIER RESILIENT TO ADVERSARIAL ATTACKS, FAULT INJECTIONS AND CONCEPT DRIFT – MODEL ARCHITECTURE AND TRAINING ALGORITHM","authors":"V. Moskalenko, A. Moskalenko, A. Korobov, M. O. Zaretsky","doi":"10.15588/1607-3274-2022-3-9","DOIUrl":"https://doi.org/10.15588/1607-3274-2022-3-9","url":null,"abstract":"Context. The problem of image classification algorithms vulnerability to destructive perturbations has not yet been definitively resolved and is quite relevant for safety-critical applications. Therefore, object of research is the process of training and inference for image classifier that functioning under influences of destructive perturbations. The subjects of the research are model architecture and training algorithm of image classifier that provide resilience to adversarial attacks, fault injection attacks and concept drift. \u0000Objective. Stated research goal is to develop effective model architecture and training algorithm that provide resilience to adversarial attacks, fault injections and concept drift. \u0000Method. New training algorithm which combines self-knowledge distillation, information measure maximization, class distribution compactness and interclass gap maximization, data compression based on discretization of feature representation and semi-supervised learning based on consistency regularization is proposed. \u0000Results. The model architecture and training algorithm of image classifier were developed. The obtained classifier was tested on the Cifar10 dataset to evaluate its resilience over an interval of 200 mini-batches with a training and test size of mini-batch equals to 128 examples for such perturbations: adversarial black-box L∞-attacks with perturbation levels equal to 1, 3, 5 and 10; inversion of one randomly selected bit in a tensor for 10%, 30%, 50% and 60% randomly selected tensors; addition of one new class; real concept drift between a pair of classes. The effect of the feature space dimensionality on the value of the information criterion of the model performance without perturbations and the value of the integral metric of resilience during the exposure to perturbations is considered. \u0000Conclusions. The proposed model architecture and learning algorithm provide absorption of part of the disturbing influence, graceful degradation due to hierarchical classes and adaptive computation, and fast adaptation on a limited amount of labeled data. It is shown that adaptive computation saves up to 40% of resources due to early decision-making in the lower sections of the model, but perturbing influence leads to slowing down, which can be considered as graceful degradation. A multi-section structure trained using knowledge self-distillation principles has been shown to provide more than 5% improvement in the value of the integral mectric of resilience compared to an architecture where the decision is made on the last layer of the model. It is observed that the dimensionality of the feature space noticeably affects the resilience to adversarial attacks and can be chosen as a tradeoff between resilience to perturbations and efficiency without perturbations.","PeriodicalId":43783,"journal":{"name":"Radio Electronics Computer Science Control","volume":"12 1","pages":""},"PeriodicalIF":0.5,"publicationDate":"2022-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76467661","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
OUTLIER DETECTION TECHNIQUE FOR HETEROGENEOUS DATA USING TRIMMED-MEAN ROBUST ESTIMATORS 采用裁剪均值鲁棒估计的异质数据离群点检测技术
IF 0.5 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2022-10-16 DOI: 10.15588/1607-3274-2022-3-5
A. Shved, Yevhen Davydenko
Context. Fortunately, the most commonly used in parametric statistics assumptions such as such as normality, linearity, independence, are not always fulfilled in real practice. The main reason for this is the appearance of observations in data samples that differ from the bulk of the data, as a result of which the sample becomes heterogeneous. The application in such conditions of generally accepted estimation procedures, for example, the sample mean, entails the bias increasing and the effectiveness decreasing of the estimates obtained. This, in turn, raises the problem of finding possible solutions to the problem of processing data sets that include outliers, especially in small samples. The object of the study is the process of detecting and excluding anomalous objects from the heterogeneous data sets. Objective. The goal of the work is to develop a procedure for anomaly detection in heterogeneous data sets, and the rationale for using a number of trimmed-mean robust estimators as a statistical measure of the location parameter of distorted parametric distribution models. Method. The problems of analysis (processing) of heterogeneous data containing outliers, sharply distinguished, suspicious observations are considered. The possibilities of using robust estimation methods for processing heterogeneous data have been analyzed. A procedure for identification and extraction of outliers caused by measurement errors, hidden equipment defects, experimental conditions, etc. has been proposed. The proposed approach is based on the procedure of symmetric and asymmetric truncation of the ranked set obtained from the initial sample of measurement data, based on the methods of robust statistics. For a reasonable choice of the value of the truncation coefficient, it is proposed to use adaptive robust procedures. Observations that fell into the zone of smallest and lowest ordinal statistics are considered outliers. Results. The proposed approach allows, in contrast to the traditional criteria for identifying outlying observations, such as the Smirnov (Grubbs) criterion, the Dixon criterion, etc., to split the analyzed set of data into a homogeneous component and identify the set of outlying observations, assuming that their share in the total set of analyzed data is unknown. Conclusions. The article proposes the use of robust statistics methods for the formation of supposed zones containing homogeneous and outlying observations in the ranked set, built on the basis of the initial sample of the analyzed data. It is proposed to use a complex of adaptive robust procedures to establish the expected truncation levels that form the zones of outlying observations in the region of the lowest and smallest order statistics of the ranked dataset. The final level of truncation of the ranked dataset is refined on the basis of existing criteria that allow checking the boundary observations (minimum and maximum) for outliers.
上下文。幸运的是,参数统计中最常用的假设,如正态性、线性、独立性,在实际实践中并不总是满足。造成这种情况的主要原因是数据样本中的观测值与大部分数据不同,因此样本变得异构。在这种情况下,应用普遍接受的估计程序,例如样本均值,意味着得到的估计的偏差增加和有效性降低。这反过来又提出了一个问题,即寻找可能的解决方案来处理包含异常值的数据集问题,特别是在小样本中。研究的对象是从异构数据集中检测和排除异常对象的过程。目标。这项工作的目标是开发一种在异构数据集中进行异常检测的程序,以及使用一些修剪平均鲁棒估计器作为扭曲参数分布模型的位置参数的统计度量的基本原理。方法。分析(处理)异构数据的问题,包括异常值,明显区分,可疑的观察。分析了使用鲁棒估计方法处理异构数据的可能性。提出了由测量误差、设备隐藏缺陷、实验条件等引起的异常值的识别和提取方法。该方法基于鲁棒统计方法,对测量数据的初始样本进行对称和非对称截断排序集。为了合理选择截断系数的取值,提出采用自适应鲁棒程序。落入最小和最低序数统计区域的观测值被认为是异常值。结果。与传统的识别离群观测值的标准(如Smirnov (Grubbs)准则、Dixon准则等)相比,该方法允许将分析的数据集分割成一个同次成分,并识别离群观测值集,假设它们在分析数据的总集合中的份额是未知的。结论。本文提出使用稳健的统计方法,在分析数据的初始样本的基础上,在排名集中形成包含均匀和离群观测的假设区域。建议使用自适应鲁棒程序的复合体来建立期望截断水平,这些截断水平形成了排序数据集的最低和最小阶统计量区域的外围观测区域。排序数据集的最终截断级别在现有标准的基础上进行改进,这些标准允许检查异常值的边界观测值(最小值和最大值)。
{"title":"OUTLIER DETECTION TECHNIQUE FOR HETEROGENEOUS DATA USING TRIMMED-MEAN ROBUST ESTIMATORS","authors":"A. Shved, Yevhen Davydenko","doi":"10.15588/1607-3274-2022-3-5","DOIUrl":"https://doi.org/10.15588/1607-3274-2022-3-5","url":null,"abstract":"Context. Fortunately, the most commonly used in parametric statistics assumptions such as such as normality, linearity, independence, are not always fulfilled in real practice. The main reason for this is the appearance of observations in data samples that differ from the bulk of the data, as a result of which the sample becomes heterogeneous. The application in such conditions of generally accepted estimation procedures, for example, the sample mean, entails the bias increasing and the effectiveness decreasing of the estimates obtained. This, in turn, raises the problem of finding possible solutions to the problem of processing data sets that include outliers, especially in small samples. The object of the study is the process of detecting and excluding anomalous objects from the heterogeneous data sets. \u0000Objective. The goal of the work is to develop a procedure for anomaly detection in heterogeneous data sets, and the rationale for using a number of trimmed-mean robust estimators as a statistical measure of the location parameter of distorted parametric distribution models. \u0000Method. The problems of analysis (processing) of heterogeneous data containing outliers, sharply distinguished, suspicious observations are considered. The possibilities of using robust estimation methods for processing heterogeneous data have been analyzed. A procedure for identification and extraction of outliers caused by measurement errors, hidden equipment defects, experimental conditions, etc. has been proposed. The proposed approach is based on the procedure of symmetric and asymmetric truncation of the ranked set obtained from the initial sample of measurement data, based on the methods of robust statistics. For a reasonable choice of the value of the truncation coefficient, it is proposed to use adaptive robust procedures. Observations that fell into the zone of smallest and lowest ordinal statistics are considered outliers. \u0000Results. The proposed approach allows, in contrast to the traditional criteria for identifying outlying observations, such as the Smirnov (Grubbs) criterion, the Dixon criterion, etc., to split the analyzed set of data into a homogeneous component and identify the set of outlying observations, assuming that their share in the total set of analyzed data is unknown. \u0000Conclusions. The article proposes the use of robust statistics methods for the formation of supposed zones containing homogeneous and outlying observations in the ranked set, built on the basis of the initial sample of the analyzed data. It is proposed to use a complex of adaptive robust procedures to establish the expected truncation levels that form the zones of outlying observations in the region of the lowest and smallest order statistics of the ranked dataset. The final level of truncation of the ranked dataset is refined on the basis of existing criteria that allow checking the boundary observations (minimum and maximum) for outliers.","PeriodicalId":43783,"journal":{"name":"Radio Electronics Computer Science Control","volume":"35 1","pages":""},"PeriodicalIF":0.5,"publicationDate":"2022-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82524507","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
KOLMOGOROV-WIENER FILTER FOR CONTINUOUS TRAFFIC PREDICTION IN THE GFSD MODEL GFSD模型中连续交通预测的Kolmogorov-wiener滤波器
IF 0.5 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2022-10-01 DOI: 10.15588/1607-3274-2022-3-3
V. Gorev, A. Gusev, V. Korniienko
Context. We investigate the Kolmogorov-Wiener filter weight function for the prediction of continuous stationary telecommunication traffic in the GFSD (Gaussian fractional sum-difference) model. Objective. The aim of the work is to obtain an approximate solution for the corresponding weight function and to illustrate the convergence of the truncated polynomial expansion method used in this paper. Method. The truncated polynomial expansion method is used for the obtaining of an approximate solution for the KolmogorovWiener weight function under consideration. In this paper we used the corresponding method on the basis of the Chebyshev polynomials of the first kind orthogonal on the time interval on which the filter input data are given. It is expected that the results based on other polynomial sets will be similar to the results obtained in this paper. Results. The weight function is investigated in the approximations up to the eighteen-polynomial one. It is shown that approximations of rather large numbers of polynomials lead to a good coincidence of the left-hand side and the right-hand side of the Wiener-Hopf integral equation. The quality of the coincidence is illustrated by the calculation of the corresponding MAPE errors. Conclusions. The paper is devoted to the theoretical construction of the Kolmogorov-Wiener filter for the prediction of continuous stationary telecommunication traffic in the GFSD model. The traffic correlation function in the framework of the GFSD model is a positively defined one, which guarantees the convergence of the truncated polynomial expansion method. The corresponding weight function is obtained in the approximations up to the eighteen-polynomial one. The convergence of the method is illustrated by the calculation of the MAPE errors of misalignment of the left-hand side and the right-hand side of the Wiener-Hopf integral equation under consideration. The results of the paper may be applied to practical traffic prediction in telecommunication systems with data packet transfer.
上下文。我们研究了高斯分数阶和差模型中用于预测连续平稳通信流量的Kolmogorov-Wiener滤波权函数。目标。本文的目的是得到相应权函数的近似解,并说明本文所使用的截断多项式展开方法的收敛性。方法。利用截断多项式展开法求出所考虑的KolmogorovWiener权函数的近似解。本文利用第一类切比雪夫多项式在给定滤波器输入数据的时间区间上的正交,给出了相应的方法。期望基于其他多项式集的结果与本文的结果相似。结果。对权函数进行了直到18次多项式的近似研究。结果表明,对相当多的多项式进行近似,可以使维纳-霍普夫积分方程的左边和右边很好地重合。通过计算相应的MAPE误差来说明重合的质量。结论。本文研究了在GFSD模型中预测连续固定通信流量的Kolmogorov-Wiener滤波器的理论构造。GFSD模型框架下的交通相关函数是一个正定义函数,保证了截断多项式展开方法的收敛性。相应的权函数在逼近到18次多项式时得到。通过计算所考虑的Wiener-Hopf积分方程的左右两侧不对准的MAPE误差,说明了该方法的收敛性。本文的研究结果可用于具有数据包传输的通信系统的实际流量预测。
{"title":"KOLMOGOROV-WIENER FILTER FOR CONTINUOUS TRAFFIC PREDICTION IN THE GFSD MODEL","authors":"V. Gorev, A. Gusev, V. Korniienko","doi":"10.15588/1607-3274-2022-3-3","DOIUrl":"https://doi.org/10.15588/1607-3274-2022-3-3","url":null,"abstract":"Context. We investigate the Kolmogorov-Wiener filter weight function for the prediction of continuous stationary telecommunication traffic in the GFSD (Gaussian fractional sum-difference) model. \u0000Objective. The aim of the work is to obtain an approximate solution for the corresponding weight function and to illustrate the convergence of the truncated polynomial expansion method used in this paper. \u0000Method. The truncated polynomial expansion method is used for the obtaining of an approximate solution for the KolmogorovWiener weight function under consideration. In this paper we used the corresponding method on the basis of the Chebyshev polynomials of the first kind orthogonal on the time interval on which the filter input data are given. It is expected that the results based on other polynomial sets will be similar to the results obtained in this paper. \u0000Results. The weight function is investigated in the approximations up to the eighteen-polynomial one. It is shown that approximations of rather large numbers of polynomials lead to a good coincidence of the left-hand side and the right-hand side of the Wiener-Hopf integral equation. The quality of the coincidence is illustrated by the calculation of the corresponding MAPE errors. \u0000Conclusions. The paper is devoted to the theoretical construction of the Kolmogorov-Wiener filter for the prediction of continuous stationary telecommunication traffic in the GFSD model. The traffic correlation function in the framework of the GFSD model is a positively defined one, which guarantees the convergence of the truncated polynomial expansion method. The corresponding weight function is obtained in the approximations up to the eighteen-polynomial one. The convergence of the method is illustrated by the calculation of the MAPE errors of misalignment of the left-hand side and the right-hand side of the Wiener-Hopf integral equation under consideration. The results of the paper may be applied to practical traffic prediction in telecommunication systems with data packet transfer.","PeriodicalId":43783,"journal":{"name":"Radio Electronics Computer Science Control","volume":"15 1","pages":""},"PeriodicalIF":0.5,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82308514","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
HOMOGENEOUS PLANS OF MULTI-FACTORY EXPERIMENTS ON QUASI-RANDOM R-ROBERTS SEQUENCES FOR SURROGATE MODELING IN A VORTEX STYLE STRUCTUROSCOPY 准随机r-roberts序列的多工厂实验齐次计划在旋涡式结构镜中进行代理建模
IF 0.5 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2022-10-01 DOI: 10.15588/1607-3274-2022-3-2
V. Galchenko, M. D. Koshevoy, R. Trembovetskaya
Context. The article is devoted to the creation of multifactorial experimental plans based on quasi-random recursive Roberts Rsequences. The object of the research is the process of creating computer-aided experimental design plans. The aim of the article is to create multifactorial, namely six- and seven-factor, uniform plans of experiments with low discrepancies, study of their projection properties and demonstrate their use on the example of surrogate modeling in eddy current structuroscopy. Method. An iterative method of unit hypercube even filling with reference points was used for constructing multidimensional experimental plans. It provides acceptable indicators of homogeneity and is realized on the basis of quasi-random nonparametric additive recursive Roberts R-sequences using irrational numbers, which, in turn, are obtained on the basis of the generalized Fibonacci sequence. The criterion for plans perfection is the assessment of homogeneity in terms of differences invariant with respect to the rotation of coordinates and re-marking and ordering of factors and which quantitatively characterize the deviation of the generated distribution from the ideal uniform. Results. Six- and seven-factor computer uniform experimental plans have been created for cataloging, which are characterized by low discrepancies and sufficiently high-quality projection properties. The tendency, which had been previously proved in the authors' research, for preserving these experimental plans characteristics in multidimensional factor spaces, which is observed with increasing number of plan points, has been confirmed. The evaluation of the quality of the created experimental plans is carried out both by visual analysis of the scattering matrix of all two-dimensional projections and by quantitative indicators of heterogeneity of the set of vectors that form the plan, namely centered and cyclic discrepancies. The example of the initial stage of creating a surrogate model to solve the problem of identifying profiles of electrophysical parameters in eddy current structuroscopy shows certain features of the application for created plans, in particular the transition from the plan for a unit hypercube to the plan in real factor space in the form of a hyperparallelepiped, which does not significantly affect its characteristics of homogeneity of the distribution of points. Conclusions. For the first time, the problem of creating six- and seven-factor uniform plans of experiments with low rates of centered and cyclic discrepancies based on R-sequences of Roberts was solved. The projection properties of the created experimental plans for different number of points were investigated. The method of constructing multidimensional computer plans of experiments taking into account the peculiarities of eddy current structuroscopy was improved. The use of six-dimensional experimental plans on the example of surrogate modeling in eddy current structuroscopy was demonstrated. Th
上下文。本文致力于基于准随机递归Roberts序列的多因子实验计划的创建。本研究的对象是计算机辅助实验设计方案的生成过程。本文的目的是创建多因素,即六因素和七因素,低差异的均匀实验计划,研究它们的投影特性,并以涡流结构镜中的代理建模为例说明它们的应用。方法。采用单位超立方体均匀填充参考点的迭代方法构建多维实验方案。它是在拟随机非参数加性递推Roberts r序列的基础上实现的,该序列是在广义Fibonacci序列的基础上得到的。计划完美的标准是根据坐标旋转和因素的重新标记和排序方面的差异不变来评估均匀性,并定量地表征生成的分布与理想均匀的偏差。结果。六因素和七因素计算机统一的编目实验计划,其特点是低差异和足够高质量的投影特性。先前在作者的研究中已经证明,随着平面点的增加,这些实验平面特征在多维因子空间中保持的趋势得到了证实。通过对所有二维投影的散射矩阵进行可视化分析,以及通过构成该计划的向量集的非均质性的定量指标(即中心差和循环差)来评估所创建的实验计划的质量。以建立替代模型解决涡流结构镜中电物理参数剖面识别问题的初始阶段为例,表明了所建平面的应用具有一定的特点,特别是从单位超立方体平面过渡到超平行六面体形式的实因子空间平面,但不明显影响其点分布均匀性的特点。结论。第一次解决了基于Roberts r序列的低中心差率和循环差率的六因子和七因子均匀实验方案的问题。研究了所创建的实验平面在不同点数下的投影特性。改进了考虑涡流结构镜特性的多维实验计算机计划构建方法。介绍了六维实验方案在涡流结构镜模拟中的应用。研究结果可以用任何已知的近似方法来建立物理过程的替代数学模型。
{"title":"HOMOGENEOUS PLANS OF MULTI-FACTORY EXPERIMENTS ON QUASI-RANDOM R-ROBERTS SEQUENCES FOR SURROGATE MODELING IN A VORTEX STYLE STRUCTUROSCOPY","authors":"V. Galchenko, M. D. Koshevoy, R. Trembovetskaya","doi":"10.15588/1607-3274-2022-3-2","DOIUrl":"https://doi.org/10.15588/1607-3274-2022-3-2","url":null,"abstract":"Context. The article is devoted to the creation of multifactorial experimental plans based on quasi-random recursive Roberts Rsequences. The object of the research is the process of creating computer-aided experimental design plans. The aim of the article is to create multifactorial, namely six- and seven-factor, uniform plans of experiments with low discrepancies, study of their projection properties and demonstrate their use on the example of surrogate modeling in eddy current structuroscopy. \u0000Method. An iterative method of unit hypercube even filling with reference points was used for constructing multidimensional experimental plans. It provides acceptable indicators of homogeneity and is realized on the basis of quasi-random nonparametric additive recursive Roberts R-sequences using irrational numbers, which, in turn, are obtained on the basis of the generalized Fibonacci sequence. The criterion for plans perfection is the assessment of homogeneity in terms of differences invariant with respect to the rotation of coordinates and re-marking and ordering of factors and which quantitatively characterize the deviation of the generated distribution from the ideal uniform. \u0000Results. Six- and seven-factor computer uniform experimental plans have been created for cataloging, which are characterized by low discrepancies and sufficiently high-quality projection properties. The tendency, which had been previously proved in the authors' research, for preserving these experimental plans characteristics in multidimensional factor spaces, which is observed with increasing number of plan points, has been confirmed. The evaluation of the quality of the created experimental plans is carried out both by visual analysis of the scattering matrix of all two-dimensional projections and by quantitative indicators of heterogeneity of the set of vectors that form the plan, namely centered and cyclic discrepancies. \u0000The example of the initial stage of creating a surrogate model to solve the problem of identifying profiles of electrophysical parameters in eddy current structuroscopy shows certain features of the application for created plans, in particular the transition from the plan for a unit hypercube to the plan in real factor space in the form of a hyperparallelepiped, which does not significantly affect its characteristics of homogeneity of the distribution of points. \u0000Conclusions. For the first time, the problem of creating six- and seven-factor uniform plans of experiments with low rates of centered and cyclic discrepancies based on R-sequences of Roberts was solved. The projection properties of the created experimental plans for different number of points were investigated. The method of constructing multidimensional computer plans of experiments taking into account the peculiarities of eddy current structuroscopy was improved. The use of six-dimensional experimental plans on the example of surrogate modeling in eddy current structuroscopy was demonstrated. Th","PeriodicalId":43783,"journal":{"name":"Radio Electronics Computer Science Control","volume":"18 1","pages":""},"PeriodicalIF":0.5,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77989803","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
ANALYSIS OF METHODS FOR AUTOMATED RESEARCH OF DC VOLTAGE CONVERTERS OF MODULAR STRUCTURE 模块化结构直流电压变换器自动化研究方法分析
IF 0.5 Q4 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Pub Date : 2022-10-01 DOI: 10.15588/1607-3274-2022-3-1
R. Kharchenko, A. V. Kochetkov, V. Mikhaylenko
Context. DC voltage converters (DCV) are part of modern power supply systems (PSS) and power supply ensuring the operation of electronic and radio devices, telecommunication systems and communication and to a large extent determine their power consumption, reliability, time of readiness for operation, weight, size and cost indicators. Even though there are a large number of different software packages used in engineering practice for the study and design of radio engineering devices, such computer-aided design (CAD) systems and virtual computer simulation of electronic circuits have some limitations that do not allow to quickly carry out the entire complex of studies of DCV required for the analysis of electrical processes in various operating modes. Objective. In this section, the goal is to select the most suitable methods and algorithms that allow the development of software necessary for solving the problems of research and analysis of electrical processes for select energy parameters of the DCV of a modular structure in a separate power channel (PWC). Method. The paper proposes a method that consists in using mathematical models describing electrical processes in DC voltage converters and creating, on the basis of the developed calculation algorithms, specialized software for the automated study of electrical processes in the DCV of a modular structure using a computer. Results. The paper discusses the main methods of automated research of radio engineering devices, which can be used to analyze the electrical processes of pulsed DC voltage converters of a modular structure. Algorithms of calculation are given and, as an example, some results of automated research obtained using this method. Conclusions. The analysis of the known methods of automated research of DC voltage converters of modular structure is carried out. Their advantages and disadvantages are given. It is shown that the most suitable method is based on the use of mathematical models describing electrical processes in DC voltage converters of this type. On the basis of the mathematical models presented in the second section of the work, algorithms and specialized software have been developed that allow them to be widely used in the automated research and design of modular-structured DC voltage converters.
上下文。直流电压变换器(DCV)是现代供电系统(PSS)的一部分,是保证电子和无线电设备、电信系统和通信运行的电源,并在很大程度上决定了它们的功耗、可靠性、运行准备时间、重量、尺寸和成本指标。尽管在工程实践中有大量不同的软件包用于无线电工程设备的研究和设计,但这些计算机辅助设计(CAD)系统和电子电路的虚拟计算机仿真存在一些局限性,无法快速执行分析各种工作模式下电气过程所需的DCV的整个复杂研究。目标。在本节中,目标是选择最合适的方法和算法,以允许开发必要的软件,以解决研究和分析在单独电源通道(PWC)中模块化结构的DCV的选择能量参数的电气过程的问题。方法。本文提出了一种用数学模型描述直流电压变换器的电过程的方法,并在已开发的计算算法的基础上,利用计算机创建用于模块化结构直流电压变换器电过程自动化研究的专用软件。结果。本文讨论了无线电工程器件自动化研究的主要方法,可用于分析模块化结构的脉冲直流电压变换器的电气过程。给出了计算算法,并举例说明了用该方法进行自动化研究的一些结果。结论。对模块化结构直流电压变换器自动化研究的已知方法进行了分析。给出了它们的优缺点。结果表明,最合适的方法是利用数学模型来描述这种类型的直流电压变换器的电气过程。在第二部分提出的数学模型的基础上,开发了算法和专用软件,使其能够广泛应用于模块化结构直流电压变换器的自动化研究和设计。
{"title":"ANALYSIS OF METHODS FOR AUTOMATED RESEARCH OF DC VOLTAGE CONVERTERS OF MODULAR STRUCTURE","authors":"R. Kharchenko, A. V. Kochetkov, V. Mikhaylenko","doi":"10.15588/1607-3274-2022-3-1","DOIUrl":"https://doi.org/10.15588/1607-3274-2022-3-1","url":null,"abstract":"Context. DC voltage converters (DCV) are part of modern power supply systems (PSS) and power supply ensuring the operation of electronic and radio devices, telecommunication systems and communication and to a large extent determine their power consumption, reliability, time of readiness for operation, weight, size and cost indicators. Even though there are a large number of different software packages used in engineering practice for the study and design of radio engineering devices, such computer-aided design (CAD) systems and virtual computer simulation of electronic circuits have some limitations that do not allow to quickly carry out the entire complex of studies of DCV required for the analysis of electrical processes in various operating modes. \u0000Objective. In this section, the goal is to select the most suitable methods and algorithms that allow the development of software necessary for solving the problems of research and analysis of electrical processes for select energy parameters of the DCV of a modular structure in a separate power channel (PWC). \u0000Method. The paper proposes a method that consists in using mathematical models describing electrical processes in DC voltage converters and creating, on the basis of the developed calculation algorithms, specialized software for the automated study of electrical processes in the DCV of a modular structure using a computer. \u0000Results. The paper discusses the main methods of automated research of radio engineering devices, which can be used to analyze the electrical processes of pulsed DC voltage converters of a modular structure. Algorithms of calculation are given and, as an example, some results of automated research obtained using this method. \u0000Conclusions. The analysis of the known methods of automated research of DC voltage converters of modular structure is carried out. Their advantages and disadvantages are given. It is shown that the most suitable method is based on the use of mathematical models describing electrical processes in DC voltage converters of this type. On the basis of the mathematical models presented in the second section of the work, algorithms and specialized software have been developed that allow them to be widely used in the automated research and design of modular-structured DC voltage converters.","PeriodicalId":43783,"journal":{"name":"Radio Electronics Computer Science Control","volume":"33 1","pages":""},"PeriodicalIF":0.5,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78063223","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Radio Electronics Computer Science Control
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1