Pub Date : 2022-12-27DOI: 10.31799/1684-8853-2022-6-2-9
N. Vassiliev, V. Duzhin, A. Kuzmin
Introduction: The RSK algorithm establishes an equivalence of finite sequences of elements of linearly ordered sets and pairs of Young tableaux P and Q of the same shape. Of particular interest is the study of the asymptotic limit, i. e., the limit shape of the so-called bumping routes formed by the boxes of tableau P affected in a single iteration of the RSK algorithm. The exact formulae for these limit shapes were obtained by D. Romik and P. Śniady in 2016. However, the problem of investigating the dynamics of the approach of bumping routes to their limit shapes remains insufficiently studied. Purpose: To study the dynamics of distances between the bumping routes and their limit shapes in Young tableaux with the help of computer experiments. Results: We have obtained a large number of experimental bumping routes through a series of computer experiments for Young tableaux P of sizes up to 4·106, filled with real numbers in the range [0, 1] and sets of inserted values α Î [0.1, 0.15, … , 0.85]. We have compared these bumping routes in the L2 metric with the corresponding limit shapes and have calculated the average distances and variances of their deviations from the limit shapes. We present an empirical formula for the rate of approach of discretized bumping routes to their limit shapes. Also, the experimental parameters of the normal distributions of the deviations of the bumping routes are obtained for various input values.
{"title":"Modeling of bumping routes in the RSK algorithm and analysis of their approach to limit shapes","authors":"N. Vassiliev, V. Duzhin, A. Kuzmin","doi":"10.31799/1684-8853-2022-6-2-9","DOIUrl":"https://doi.org/10.31799/1684-8853-2022-6-2-9","url":null,"abstract":"Introduction: The RSK algorithm establishes an equivalence of finite sequences of elements of linearly ordered sets and pairs of Young tableaux P and Q of the same shape. Of particular interest is the study of the asymptotic limit, i. e., the limit shape of the so-called bumping routes formed by the boxes of tableau P affected in a single iteration of the RSK algorithm. The exact formulae for these limit shapes were obtained by D. Romik and P. Śniady in 2016. However, the problem of investigating the dynamics of the approach of bumping routes to their limit shapes remains insufficiently studied. Purpose: To study the dynamics of distances between the bumping routes and their limit shapes in Young tableaux with the help of computer experiments. Results: We have obtained a large number of experimental bumping routes through a series of computer experiments for Young tableaux P of sizes up to 4·106, filled with real numbers in the range [0, 1] and sets of inserted values α Î [0.1, 0.15, … , 0.85]. We have compared these bumping routes in the L2 metric with the corresponding limit shapes and have calculated the average distances and variances of their deviations from the limit shapes. We present an empirical formula for the rate of approach of discretized bumping routes to their limit shapes. Also, the experimental parameters of the normal distributions of the deviations of the bumping routes are obtained for various input values.","PeriodicalId":36977,"journal":{"name":"Informatsionno-Upravliaiushchie Sistemy","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44041638","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-12-27DOI: 10.31799/1684-8853-2022-6-31-40
D. Stepanov, V. Itsykson
Introduction: The standard way to check the quality of a compiler is manual testing. However, it does not allow to cover a vast diversity of programs that can be written in a target programming language. Today, in addition to manual written tests there are many automated compiler testing methods, among which fuzzing is one of the most powerful and useful. A compiler fuzzer is a tool that generates a random program in a target language and checks how the compiler works in this language. Purpose: To develop a platform for compiler fuzzing and, based on it, to develop a tool for Kotlin compiler testing. Results: We have developed Backend Bug Finder which is a platform for compiler fuzzing is. We have chosen a mutation-based approach as a method for generating random programs. First, an existing program is entered to the mutator as the input to be then transformed in some way. Mutations can be both trivial, for example, replacing arithmetic operators with others, and complex, changing the structure of the program. Next, the resulting program is fed to the input of the compiler with the following check of its operation. The developed test oracle can detect three types of errors: crashes, miscompilations, and performance degradations. If an error is detected, the test case is fed into the post-processing module, where reduction and deduplication algorithms are applied. We have developed a tool for fuzzing the Kotlin language compiler based on the platform for its approbation, which showed the applicability of the proposed approach for finding errors in modern compilers. Practical relevance: Over a year and a half of work, our tool has found thousands of different Kotlin compiler bugs, more than 200 of which were sent to the developers, and more than 80 have been fixed.
{"title":"Backend Bug Finder — a platform for effective compiler fuzzing","authors":"D. Stepanov, V. Itsykson","doi":"10.31799/1684-8853-2022-6-31-40","DOIUrl":"https://doi.org/10.31799/1684-8853-2022-6-31-40","url":null,"abstract":"Introduction: The standard way to check the quality of a compiler is manual testing. However, it does not allow to cover a vast diversity of programs that can be written in a target programming language. Today, in addition to manual written tests there are many automated compiler testing methods, among which fuzzing is one of the most powerful and useful. A compiler fuzzer is a tool that generates a random program in a target language and checks how the compiler works in this language. Purpose: To develop a platform for compiler fuzzing and, based on it, to develop a tool for Kotlin compiler testing. Results: We have developed Backend Bug Finder which is a platform for compiler fuzzing is. We have chosen a mutation-based approach as a method for generating random programs. First, an existing program is entered to the mutator as the input to be then transformed in some way. Mutations can be both trivial, for example, replacing arithmetic operators with others, and complex, changing the structure of the program. Next, the resulting program is fed to the input of the compiler with the following check of its operation. The developed test oracle can detect three types of errors: crashes, miscompilations, and performance degradations. If an error is detected, the test case is fed into the post-processing module, where reduction and deduplication algorithms are applied. We have developed a tool for fuzzing the Kotlin language compiler based on the platform for its approbation, which showed the applicability of the proposed approach for finding errors in modern compilers. Practical relevance: Over a year and a half of work, our tool has found thousands of different Kotlin compiler bugs, more than 200 of which were sent to the developers, and more than 80 have been fixed.","PeriodicalId":36977,"journal":{"name":"Informatsionno-Upravliaiushchie Sistemy","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48937150","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-12-27DOI: 10.31799/1684-8853-2022-6-41-52
A. Ovchinnikov, A. Veresova, A. Fominykh
Introduction: In modern systems for communication, data storage and processing the error-correction capability of codes are estimated for memoryless channels. In real channels the noise is correlated, which leads to grouping error in bursts. A traditional method to fight this phenomenon is channel decorrelation, which does not allow developing of coding schemes, mostly utilizing the channel capacity. Thus the development of bursts decoding algorithms for arbitrary linear codes is the actual task. Purpose: To develop a single error burst decoding algorithm for linear codes, to estimate the decoding error probability and computational complexity. Results: Two approaches are proposed to burst error correction. The first one is based on combining the window sliding modification of well-known bit-flipping algorithm with preliminary analysis of the structure of parity check matrix. The second one is based on the recursive procedure of constructing the sequence of certain events which, in the worst case, performs the exhaustive search of error bursts, but in many cases the search may be significantly decreased by using the proposed heuristics. The proposed recursive decoding algorithm allows a guaranteed correction of any single error bursts within burst-correction capability of the code, and in many cases beyond the burst-correction capability. The complexity of this algorithm is significantly lower than that of a bit flipping algorithm if the parity-check matrix of the code is sparse enough. An alternative hybrid decoding algorithm is proposed utilizing the bit-flipping approach and showing the error probability and completion time comparable to the recursive algorithm, however, in this case the possibility of a guaranteed burst correction hardly can be proved. Practical relevance: The proposed decoding methods may be used in modern and perspective communication systems, allowing energy saving and increasing reliability of data transmission by better error performance and computational complexity.
{"title":"Decoding of linear codes for single error bursts correction based on the determination of certain events","authors":"A. Ovchinnikov, A. Veresova, A. Fominykh","doi":"10.31799/1684-8853-2022-6-41-52","DOIUrl":"https://doi.org/10.31799/1684-8853-2022-6-41-52","url":null,"abstract":"Introduction: In modern systems for communication, data storage and processing the error-correction capability of codes are estimated for memoryless channels. In real channels the noise is correlated, which leads to grouping error in bursts. A traditional method to fight this phenomenon is channel decorrelation, which does not allow developing of coding schemes, mostly utilizing the channel capacity. Thus the development of bursts decoding algorithms for arbitrary linear codes is the actual task. Purpose: To develop a single error burst decoding algorithm for linear codes, to estimate the decoding error probability and computational complexity. Results: Two approaches are proposed to burst error correction. The first one is based on combining the window sliding modification of well-known bit-flipping algorithm with preliminary analysis of the structure of parity check matrix. The second one is based on the recursive procedure of constructing the sequence of certain events which, in the worst case, performs the exhaustive search of error bursts, but in many cases the search may be significantly decreased by using the proposed heuristics. The proposed recursive decoding algorithm allows a guaranteed correction of any single error bursts within burst-correction capability of the code, and in many cases beyond the burst-correction capability. The complexity of this algorithm is significantly lower than that of a bit flipping algorithm if the parity-check matrix of the code is sparse enough. An alternative hybrid decoding algorithm is proposed utilizing the bit-flipping approach and showing the error probability and completion time comparable to the recursive algorithm, however, in this case the possibility of a guaranteed burst correction hardly can be proved. Practical relevance: The proposed decoding methods may be used in modern and perspective communication systems, allowing energy saving and increasing reliability of data transmission by better error performance and computational complexity.","PeriodicalId":36977,"journal":{"name":"Informatsionno-Upravliaiushchie Sistemy","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47342037","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-12-27DOI: 10.31799/1684-8853-2022-6-10-19
Yaroslav Zhurba, A. Filchenkov, A. Azarov, A. Shalyto
Introduction: We consider the problem of routing of piece cargo by a conveyor system. When moving cargo pieces, it is necessary not only to minimize the time of transportation, but also to minimize the energy spent on it. Purpose: Development of a routing algorithm that is adaptive to changes in the topology of the routing graph and is able to optimize the delivery time and the consumed energy. Results: We propose an algorithm based on multi-agent deep reinforcement learning that places agents at the vertices of a conveyor network graph and uses a new state value function. The algorithm has two tunable parameters: the length of the path along which the state value function is calculated, and the learning coefficient. Through the selection of parameters, we have revealed that the optimal values are 2 and 1, respectively. An experimental study of the algorithm using a simulation model has shown that it allows to reduce the number of collisions of moving objects to zero, demonstrates stable results for both optimized scores, and also leads to a lower energy consumption compared with the method used as a baseline. Practical relevance: The proposed algorithm can be used to reduce delivery time and energy when managing conveyor systems.
{"title":"Continuous control algorithms for conveyer belt routing based on multi-agent deep reinforcement learning","authors":"Yaroslav Zhurba, A. Filchenkov, A. Azarov, A. Shalyto","doi":"10.31799/1684-8853-2022-6-10-19","DOIUrl":"https://doi.org/10.31799/1684-8853-2022-6-10-19","url":null,"abstract":"Introduction: We consider the problem of routing of piece cargo by a conveyor system. When moving cargo pieces, it is necessary not only to minimize the time of transportation, but also to minimize the energy spent on it. Purpose: Development of a routing algorithm that is adaptive to changes in the topology of the routing graph and is able to optimize the delivery time and the consumed energy. Results: We propose an algorithm based on multi-agent deep reinforcement learning that places agents at the vertices of a conveyor network graph and uses a new state value function. The algorithm has two tunable parameters: the length of the path along which the state value function is calculated, and the learning coefficient. Through the selection of parameters, we have revealed that the optimal values are 2 and 1, respectively. An experimental study of the algorithm using a simulation model has shown that it allows to reduce the number of collisions of moving objects to zero, demonstrates stable results for both optimized scores, and also leads to a lower energy consumption compared with the method used as a baseline. Practical relevance: The proposed algorithm can be used to reduce delivery time and energy when managing conveyor systems.","PeriodicalId":36977,"journal":{"name":"Informatsionno-Upravliaiushchie Sistemy","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45049378","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-12-27DOI: 10.31799/1684-8853-2022-6-20-30
G. Voronkov, A.G. Zakoyan, Vladislav Ivanov, A. Voronkova, I. Stepanov, Ellizaveta Grakhova, V. Lyubopytov, R. Kutluyarov
Introduction: Today sensor systems based on integrated photonics devices are the most important branch of embedded information and control systems for various functions. The output characteristics of a sensor system are significantly determined by the efficiency of the interrogator. The intensity interrogator based on a microring resonator can provide a high scanning rate and sensitivity that meets the requirements of a wide range of applications. Purpose: To develop an effective sensor system composed of a refractometric sensor and an interrogator located on the same photonic integrated circuit for marker-free determination of the concentration of substances in liquids. Methods: We use the numerical simulation of electromagnetic field propagation in a waveguide system (integrated silicon waveguides on a silicon dioxide substrate) in the research. The simulation has been carried out using the Ansys Lumerical environment, the FDTD (Finite Difference Time Domain) solver. The parameters of the microring resonators were optimized to obtain the coupling coefficients between the waveguides, providing the operation in the critical coupling mode. Results: We propose the concept of a fully integrated photonic sensor system based on micro-ring add-drop resonators. A sensor based on microring resonators has been developed, which consists of two half-rings with a radius of 18 μm, connected by sections of straight waveguides 3 μm long. An interrogator represented by a microring resonator with a radius of 10 µm has been developed. According to simulation results with a broadband source, the achieved sensor sensitivity was 110 nm per refractive index change, or 1350 dB per refractive index change. We propose a technique for choosing the optimal characteristics of the sensor and interrogator targeted to improve the complete system efficiency. Practical relevance: Sensor systems based on photonic integrated circuits can meet the demand for devices characterized by low power consumption, small size, immunity to electromagnetic interference and low cost.
目前,基于集成光子器件的传感器系统是嵌入式信息与控制系统中最重要的分支,具有多种功能。传感器系统的输出特性在很大程度上取决于询问器的效率。基于微环谐振腔的强度询问器可以提供高扫描速率和灵敏度,满足广泛应用的要求。目的:研制一种在同一光子集成电路上由折射传感器和询问器组成的有效传感器系统,用于液体中物质浓度的无标记测定。方法:采用数值模拟方法研究电磁场在波导系统(二氧化硅衬底上的集成硅波导)中的传播。利用Ansys有限元环境、时域有限差分(FDTD)求解器进行了仿真。通过优化微环谐振器的参数,获得了波导间的耦合系数,从而实现了在临界耦合模式下的工作。结果:我们提出了基于微环加降谐振器的全集成光子传感器系统的概念。研制了一种基于微环谐振腔的传感器,该传感器由两个半径为18 μm的半环组成,由3 μm长的直波导连接。研制了一种以半径为10 μ m的微环谐振器为代表的询问器。根据宽带源的仿真结果,实现的传感器灵敏度为110 nm /折射率变化,或1350 dB /折射率变化。我们提出了一种选择传感器和询问器的最佳特性的技术,以提高整个系统的效率。实际意义:基于光子集成电路的传感器系统可以满足低功耗、小体积、抗电磁干扰、低成本的器件需求。
{"title":"Fully integrated optical sensor system with intensity interrogation","authors":"G. Voronkov, A.G. Zakoyan, Vladislav Ivanov, A. Voronkova, I. Stepanov, Ellizaveta Grakhova, V. Lyubopytov, R. Kutluyarov","doi":"10.31799/1684-8853-2022-6-20-30","DOIUrl":"https://doi.org/10.31799/1684-8853-2022-6-20-30","url":null,"abstract":"Introduction: Today sensor systems based on integrated photonics devices are the most important branch of embedded information and control systems for various functions. The output characteristics of a sensor system are significantly determined by the efficiency of the interrogator. The intensity interrogator based on a microring resonator can provide a high scanning rate and sensitivity that meets the requirements of a wide range of applications. Purpose: To develop an effective sensor system composed of a refractometric sensor and an interrogator located on the same photonic integrated circuit for marker-free determination of the concentration of substances in liquids. Methods: We use the numerical simulation of electromagnetic field propagation in a waveguide system (integrated silicon waveguides on a silicon dioxide substrate) in the research. The simulation has been carried out using the Ansys Lumerical environment, the FDTD (Finite Difference Time Domain) solver. The parameters of the microring resonators were optimized to obtain the coupling coefficients between the waveguides, providing the operation in the critical coupling mode. Results: We propose the concept of a fully integrated photonic sensor system based on micro-ring add-drop resonators. A sensor based on microring resonators has been developed, which consists of two half-rings with a radius of 18 μm, connected by sections of straight waveguides 3 μm long. An interrogator represented by a microring resonator with a radius of 10 µm has been developed. According to simulation results with a broadband source, the achieved sensor sensitivity was 110 nm per refractive index change, or 1350 dB per refractive index change. We propose a technique for choosing the optimal characteristics of the sensor and interrogator targeted to improve the complete system efficiency. Practical relevance: Sensor systems based on photonic integrated circuits can meet the demand for devices characterized by low power consumption, small size, immunity to electromagnetic interference and low cost.","PeriodicalId":36977,"journal":{"name":"Informatsionno-Upravliaiushchie Sistemy","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46523420","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-12-27DOI: 10.31799/1684-8853-2022-6-53-61
V. Ponomarenko, E. Lapsheva, M. Prokhorov
Introduction: The problem of delay time estimation in ring self-oscillatory time-delay systems arises in various fields of science and is of great importance in the study of real systems generating chaotic time series. Purpose: To conduct a comparative analysis of the operation of methods for the reconstruction of time-delay systems from chaotic time series in the absence and presence of additive noise. Methods: Methods for estimating the delay time according to the statistics of extrema, using the autocorrelation function and the method of order time asymmetry are used. Based on the latter method, a method is proposed that is focused on estimating the delay times in systems with two delays. Results: We carry out a comparative analysis of the operation of four methods for reconstructing the delay times in self-oscillating time-delay systems from chaotic time series using the example of Ikeda systems with one and two delay times. We demonstrate that in the absence of additive noise, the delay time estimation method based on statistics of extrema is the most accurate one for the case of time series analysis of systems with both one and two delays. In the presence of additive noise, the modified method of order time asymmetry proposed in the work in the case of the analysis of systems with one delay time works no worse than the method of the autocorrelation function and order time asymmetry. In the case of two delay times, the modified order time asymmetry method works better than others. Practical relevance: The described methods can have a practical application in estimating the delay time of self-oscillating systems, yet the level of additive noise can affect the accuracy of the estimate.
{"title":"Estimation of delay times from time series of ring self-oscillatory time-delay systems","authors":"V. Ponomarenko, E. Lapsheva, M. Prokhorov","doi":"10.31799/1684-8853-2022-6-53-61","DOIUrl":"https://doi.org/10.31799/1684-8853-2022-6-53-61","url":null,"abstract":"Introduction: The problem of delay time estimation in ring self-oscillatory time-delay systems arises in various fields of science and is of great importance in the study of real systems generating chaotic time series. Purpose: To conduct a comparative analysis of the operation of methods for the reconstruction of time-delay systems from chaotic time series in the absence and presence of additive noise. Methods: Methods for estimating the delay time according to the statistics of extrema, using the autocorrelation function and the method of order time asymmetry are used. Based on the latter method, a method is proposed that is focused on estimating the delay times in systems with two delays. Results: We carry out a comparative analysis of the operation of four methods for reconstructing the delay times in self-oscillating time-delay systems from chaotic time series using the example of Ikeda systems with one and two delay times. We demonstrate that in the absence of additive noise, the delay time estimation method based on statistics of extrema is the most accurate one for the case of time series analysis of systems with both one and two delays. In the presence of additive noise, the modified method of order time asymmetry proposed in the work in the case of the analysis of systems with one delay time works no worse than the method of the autocorrelation function and order time asymmetry. In the case of two delay times, the modified order time asymmetry method works better than others. Practical relevance: The described methods can have a practical application in estimating the delay time of self-oscillating systems, yet the level of additive noise can affect the accuracy of the estimate.","PeriodicalId":36977,"journal":{"name":"Informatsionno-Upravliaiushchie Sistemy","volume":"1 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"69413414","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-10-28DOI: 10.31799/1684-8853-2022-5-2-11
Yevgeniya Tyryshkina, S. Tumkovskiy
Introduction: rapidly growing volumes of information pose new challenges to modern data analysis technologies. Currently, based on cost and performance considerations, data processing is usually performed in cluster systems. One of the most common related operations in analytics is the joins of datasets. Join is an extremely expensive operation that is difficult to scale and increase efficiency in distributed databases or systems based on the MapReduce paradigm. Despite the fact that a lot of effort has been put into improving the performance of this operation, often the proposed methods either require fundamental changes in the MapReduce structure, or are aimed at reducing the overhead of the operation, such as balancing the load on the network. Objective: to develop an algorithm to accelerate the integration of data sets in distributed systems. Results: a review of the Apache Spark architecture and the features of distributed computing based on MapReduce is performed, typical methods for combining datasets are analyzed, the main recommendations for optimizing the operation of combining data are presented, an algorithm that allows you to speed up the special case of combining implemented in Apache Spark is presented. This algorithm uses the methods of partitioning and partial transfer of sets to the computing nodes of the cluster, in such a way as to take advantage of the merge and broadcast associations. The experimental data presented demonstrate that the method is all the more effective the larger the volume of input data. So, for 2Tb compressed data, acceleration up to ~37% was obtained in comparison with standard Spark SQL.
{"title":"Method for accelerating the joining of distributed datasets by a given criterion","authors":"Yevgeniya Tyryshkina, S. Tumkovskiy","doi":"10.31799/1684-8853-2022-5-2-11","DOIUrl":"https://doi.org/10.31799/1684-8853-2022-5-2-11","url":null,"abstract":"Introduction: rapidly growing volumes of information pose new challenges to modern data analysis technologies. Currently, based on cost and performance considerations, data processing is usually performed in cluster systems. One of the most common related operations in analytics is the joins of datasets. Join is an extremely expensive operation that is difficult to scale and increase efficiency in distributed databases or systems based on the MapReduce paradigm. Despite the fact that a lot of effort has been put into improving the performance of this operation, often the proposed methods either require fundamental changes in the MapReduce structure, or are aimed at reducing the overhead of the operation, such as balancing the load on the network. Objective: to develop an algorithm to accelerate the integration of data sets in distributed systems. Results: a review of the Apache Spark architecture and the features of distributed computing based on MapReduce is performed, typical methods for combining datasets are analyzed, the main recommendations for optimizing the operation of combining data are presented, an algorithm that allows you to speed up the special case of combining implemented in Apache Spark is presented. This algorithm uses the methods of partitioning and partial transfer of sets to the computing nodes of the cluster, in such a way as to take advantage of the merge and broadcast associations. The experimental data presented demonstrate that the method is all the more effective the larger the volume of input data. So, for 2Tb compressed data, acceleration up to ~37% was obtained in comparison with standard Spark SQL.","PeriodicalId":36977,"journal":{"name":"Informatsionno-Upravliaiushchie Sistemy","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42821585","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-10-28DOI: 10.31799/1684-8853-2022-5-32-39
A. Aderkina, A. Sinitsyn
Introduction: The modern approach to radio planning provides subway passengers with uninterrupted access to the Internet. This is achieved through the use of a special signal propagation model which calculates signal power loss during its propagation between a transmitter and a receiver on subway lines. The disadvantage of the model is the high computational complexity. Purpose: Using machine learning methods to develop an algorithm for predicting the signal power loss, the algorithm being characterized by high accuracy and low computational complexity. Results: The analysis of machine learning methods revealed that the maximum possible accuracy in solving the problem is provided by the random forest method. A data structure containing the parameters of a digital map of subway lines was developed to train the selected method and predict a signal power loss. While developing the final algorithm a number of assumptions were made, such as: the problem is solved as a classification problem, the predicted values are integers. A signal power loss prediction algorithm that does not directly use the propagation model was developed, which reduced the computational complexity and the execution time for solving radio planning problems, with high prediction accuracy maintained. Practical relevance: Due to the use of machine learning methods in developed algorithms the time for performing radio planning was reduced from several days to several hours, with accuracy preserved. This allows to process more radio planning orders or to reduce the working time for engineers to complete the same number of orders, which is a financial benefit.
{"title":"Development of an automated software tool based on machine learning methods for solving problems of radio planning in subway sections","authors":"A. Aderkina, A. Sinitsyn","doi":"10.31799/1684-8853-2022-5-32-39","DOIUrl":"https://doi.org/10.31799/1684-8853-2022-5-32-39","url":null,"abstract":"Introduction: The modern approach to radio planning provides subway passengers with uninterrupted access to the Internet. This is achieved through the use of a special signal propagation model which calculates signal power loss during its propagation between a transmitter and a receiver on subway lines. The disadvantage of the model is the high computational complexity. Purpose: Using machine learning methods to develop an algorithm for predicting the signal power loss, the algorithm being characterized by high accuracy and low computational complexity. Results: The analysis of machine learning methods revealed that the maximum possible accuracy in solving the problem is provided by the random forest method. A data structure containing the parameters of a digital map of subway lines was developed to train the selected method and predict a signal power loss. While developing the final algorithm a number of assumptions were made, such as: the problem is solved as a classification problem, the predicted values are integers. A signal power loss prediction algorithm that does not directly use the propagation model was developed, which reduced the computational complexity and the execution time for solving radio planning problems, with high prediction accuracy maintained. Practical relevance: Due to the use of machine learning methods in developed algorithms the time for performing radio planning was reduced from several days to several hours, with accuracy preserved. This allows to process more radio planning orders or to reduce the working time for engineers to complete the same number of orders, which is a financial benefit.","PeriodicalId":36977,"journal":{"name":"Informatsionno-Upravliaiushchie Sistemy","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46814473","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-10-28DOI: 10.31799/1684-8853-2022-5-40-48
A. Pimenov, I. Fedorov, S. Bezzateev
Introduction: Due to the growth in the number and variety of devices connected to the Internet, the requirements for network performance and data transmission security are increasing. Today, performance problems are usually solved through cloud, fog and edge computing, while the problem of data storage and transmission security remains relevant. One of the effective ways to solve this problem is to use blockchain technology. Purpose: Designing the architecture of a fog computing network based on blockchain technology. Results: Based on the research in the field of fog computing, the requirements for the fog computing architecture were determined, such as: autonomy, scalability, flexibility, hierarchy, security, reliability, availability, serviceability. The selected criteria for building an architecture led to the choice in favor of a private blockchain due to its higher performance compared to a public blockchain A comparative analysis of the consensus algorithms that are most often used in private blockchains was carried out and the most suitable one was chosen. Based on the requirements put forward and the results of the analysis, a fog computing architecture model based on a private blockchain was designed. The architecture consists of four elements: end devices, fog nodes, orchestration nodes, and cloud infrastructure. The blockchain includes fog nodes and orchestration nodes, which ensures the confidentiality, availability and integrity of data in the fog network. Practical relevance: Paper results can be used in the design of fog computing networks both separately and as part of 5G mobile networks.
{"title":"Fog computing architecture using blockchain technology","authors":"A. Pimenov, I. Fedorov, S. Bezzateev","doi":"10.31799/1684-8853-2022-5-40-48","DOIUrl":"https://doi.org/10.31799/1684-8853-2022-5-40-48","url":null,"abstract":"Introduction: Due to the growth in the number and variety of devices connected to the Internet, the requirements for network performance and data transmission security are increasing. Today, performance problems are usually solved through cloud, fog and edge computing, while the problem of data storage and transmission security remains relevant. One of the effective ways to solve this problem is to use blockchain technology. Purpose: Designing the architecture of a fog computing network based on blockchain technology. Results: Based on the research in the field of fog computing, the requirements for the fog computing architecture were determined, such as: autonomy, scalability, flexibility, hierarchy, security, reliability, availability, serviceability. The selected criteria for building an architecture led to the choice in favor of a private blockchain due to its higher performance compared to a public blockchain A comparative analysis of the consensus algorithms that are most often used in private blockchains was carried out and the most suitable one was chosen. Based on the requirements put forward and the results of the analysis, a fog computing architecture model based on a private blockchain was designed. The architecture consists of four elements: end devices, fog nodes, orchestration nodes, and cloud infrastructure. The blockchain includes fog nodes and orchestration nodes, which ensures the confidentiality, availability and integrity of data in the fog network. Practical relevance: Paper results can be used in the design of fog computing networks both separately and as part of 5G mobile networks.","PeriodicalId":36977,"journal":{"name":"Informatsionno-Upravliaiushchie Sistemy","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45589077","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-10-28DOI: 10.31799/1684-8853-2022-5-49-59
A. Burkov, A. Turlikov, Roman Rachugin
Introduction: Internet of Things devices are actively used within the framework of Massive Machine-Type Communication scenarios. The interaction of devices is carried out by random multiple-access algorithms with limited throughput. To improve throughput one can use orthogonal preambles in the ALOHA-type class of algorithms. Purpose: To analyze ALOHA-based algorithms using the exploration phase and to calculate the characteristics for the algorithm with and without losses with a finite number of channels. Results: We have described a system model that employs random access for data transmission over a common communication channel with the use of orthogonal preambles and exploration phase. We have obtained a formula for numerical calculation of the throughput of an algorithm channel with losses with an infinite number of preambles and a given finite number of channels. The calculation results for several values of the number of independent channels are presented. A modification of the algorithm using the exploration phase and repeated transmissions is proposed and described. The system in question can work without losses. For this system, we have given the analysis of the maximum input throughput up to which the system operates stably. Also, the average delay values for the algorithm that were obtained by simulation modeling are shown. By reducing the number of available preambles, the results obtained can be used as an upper bound on the system throughput. Practical relevance: The results obtained allow to assess the potential for improving the throughput of random multiple-access systems in 6G networks through the application of the exploration phase.
{"title":"Analyzing and stabilizing multichannel ALOHA with the use of the preamble-based exploration phase","authors":"A. Burkov, A. Turlikov, Roman Rachugin","doi":"10.31799/1684-8853-2022-5-49-59","DOIUrl":"https://doi.org/10.31799/1684-8853-2022-5-49-59","url":null,"abstract":"Introduction: Internet of Things devices are actively used within the framework of Massive Machine-Type Communication scenarios. The interaction of devices is carried out by random multiple-access algorithms with limited throughput. To improve throughput one can use orthogonal preambles in the ALOHA-type class of algorithms. Purpose: To analyze ALOHA-based algorithms using the exploration phase and to calculate the characteristics for the algorithm with and without losses with a finite number of channels. Results: We have described a system model that employs random access for data transmission over a common communication channel with the use of orthogonal preambles and exploration phase. We have obtained a formula for numerical calculation of the throughput of an algorithm channel with losses with an infinite number of preambles and a given finite number of channels. The calculation results for several values of the number of independent channels are presented. A modification of the algorithm using the exploration phase and repeated transmissions is proposed and described. The system in question can work without losses. For this system, we have given the analysis of the maximum input throughput up to which the system operates stably. Also, the average delay values for the algorithm that were obtained by simulation modeling are shown. By reducing the number of available preambles, the results obtained can be used as an upper bound on the system throughput. Practical relevance: The results obtained allow to assess the potential for improving the throughput of random multiple-access systems in 6G networks through the application of the exploration phase.","PeriodicalId":36977,"journal":{"name":"Informatsionno-Upravliaiushchie Sistemy","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46262656","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}