Pub Date : 2020-05-11DOI: 10.15625/1813-9663/36/2/14578
Cong Thanh Bui, Loi Cao Van, Minh Hoang, Quang Uy Nguyen
The rapid development of the Internet and the wide spread of its applications has affected many aspects of our life. However, this development also makes the cyberspace more vulnerable to various attacks. Thus, detecting and preventing these attacks are crucial for the next development of the Internet and its services. Recently, machine learning methods have been widely adopted in detecting network attacks. Among many machine learning methods, AutoEncoders (AEs) are known as the state-of-the-art techniques for network anomaly detection. Although, AEs have been successfully applied to detect many types of attacks, it is often unable to detect some difficult attacks that attempt to mimic the normal network traffic. In order to handle this issue, we propose a new model based on AutoEncoder called Double-Shrink AutoEncoder (DSAE). DSAE put more shrinkage on the normal data in the middle hidden layer. This helps to pull out some anomalies that are very similar to normal data. DSAE are evaluated on six well-known network attacks datasets. The experimental results show that our model performs competitively to the state-of-the-art model, and often out-performs this model on the attacks group that is difficult for the previous methods.
{"title":"A DOUBLE-SHRINK AUTOENCODER FOR NETWORK ANOMALY DETECTION","authors":"Cong Thanh Bui, Loi Cao Van, Minh Hoang, Quang Uy Nguyen","doi":"10.15625/1813-9663/36/2/14578","DOIUrl":"https://doi.org/10.15625/1813-9663/36/2/14578","url":null,"abstract":"The rapid development of the Internet and the wide spread of its applications has affected many aspects of our life. However, this development also makes the cyberspace more vulnerable to various attacks. Thus, detecting and preventing these attacks are crucial for the next development of the Internet and its services. Recently, machine learning methods have been widely adopted in detecting network attacks. Among many machine learning methods, AutoEncoders (AEs) are known as the state-of-the-art techniques for network anomaly detection. Although, AEs have been successfully applied to detect many types of attacks, it is often unable to detect some difficult attacks that attempt to mimic the normal network traffic. In order to handle this issue, we propose a new model based on AutoEncoder called Double-Shrink AutoEncoder (DSAE). DSAE put more shrinkage on the normal data in the middle hidden layer. This helps to pull out some anomalies that are very similar to normal data. DSAE are evaluated on six well-known network attacks datasets. The experimental results show that our model performs competitively to the state-of-the-art model, and often out-performs this model on the attacks group that is difficult for the previous methods.","PeriodicalId":15444,"journal":{"name":"Journal of Computer Science and Cybernetics","volume":"1 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91366130","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2020-05-11DOI: 10.15625/1813-9663/36/2/14396
N. D. Hieu, N. C. Ho, Vũ Như Lân
Dealing with the time series forecasting problem attracts much attention from the fuzzy community. Many models and methods have been proposed in the literature since the publication of the study by Song and Chissom in 1993, in which they proposed fuzzy time series together with its fuzzy forecasting model for time series data and the fuzzy formalism to handle their uncertainty. Unfortunately, the proposed method to calculate this fuzzy model was very complex. Then, in 1996, Chen proposed an efficient method to reduce the computational complexity of the mentioned formalism. Hwang et al. in 1998 proposed a new fuzzy time series forecasting model, which deals with the variations of historical data instead of these historical data themselves. Though fuzzy sets are concepts inspired by fuzzy linguistic information, there is no formal bridge to connect the fuzzy sets and the inherent quantitative semantics of linguistic words. This study proposes the so-called linguistic time series, in which words with their own semantics are used instead of fuzzy sets. By this, forecasting linguistic logical relationships can be established based on the time series variations and this is clearly useful for human users. The effect of the proposed model is justified by applying the proposed model to forecast student enrollment historical data.
{"title":"ENROLLMENT FORECASTING BASED ON LINGUISTIC TIME SERIES","authors":"N. D. Hieu, N. C. Ho, Vũ Như Lân","doi":"10.15625/1813-9663/36/2/14396","DOIUrl":"https://doi.org/10.15625/1813-9663/36/2/14396","url":null,"abstract":"Dealing with the time series forecasting problem attracts much attention from the fuzzy community. Many models and methods have been proposed in the literature since the publication of the study by Song and Chissom in 1993, in which they proposed fuzzy time series together with its fuzzy forecasting model for time series data and the fuzzy formalism to handle their uncertainty. Unfortunately, the proposed method to calculate this fuzzy model was very complex. Then, in 1996, Chen proposed an efficient method to reduce the computational complexity of the mentioned formalism. Hwang et al. in 1998 proposed a new fuzzy time series forecasting model, which deals with the variations of historical data instead of these historical data themselves. Though fuzzy sets are concepts inspired by fuzzy linguistic information, there is no formal bridge to connect the fuzzy sets and the inherent quantitative semantics of linguistic words. This study proposes the so-called linguistic time series, in which words with their own semantics are used instead of fuzzy sets. By this, forecasting linguistic logical relationships can be established based on the time series variations and this is clearly useful for human users. The effect of the proposed model is justified by applying the proposed model to forecast student enrollment historical data.","PeriodicalId":15444,"journal":{"name":"Journal of Computer Science and Cybernetics","volume":"45 1","pages":"119-137"},"PeriodicalIF":0.0,"publicationDate":"2020-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73899589","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2020-05-11DOI: 10.15625/1813-9663/36/2/14786
Hoang Ngoc Thanh, T. Lang
The UNSW-NB15 dataset was created by the Australian Cyber Security Centre in 2015 by using the IXIA tool to extract normal behaviors and modern attacks, it includes normal data and 9 types of attacks with 49 features. Previous research results show that the detection of Fuzzers attacks in this dataset gives the lowest classification quality. This paper analyzes and evaluates the performance of using known ensemble techniques such as Bagging, AdaBoost, Stacking, Decorate, Random Forest and Voting to detect FUZZERS attacks on UNSW-NB15 dataset to create models. The experimental results show that the AdaBoost technique with the component classifiers using decision tree for the best classification quality with F −Measure is 96.76% compared to 94.16%, which is the best result by using single classifiers and 96.36% by using the Random Forest technique.
{"title":"EVALUATING EFFECTIVENESS OF ENSEMBLE CLASSIFIERS WHEN DETECTING FUZZERS ATTACKS ON THE UNSW-NB15 DATASET","authors":"Hoang Ngoc Thanh, T. Lang","doi":"10.15625/1813-9663/36/2/14786","DOIUrl":"https://doi.org/10.15625/1813-9663/36/2/14786","url":null,"abstract":"The UNSW-NB15 dataset was created by the Australian Cyber Security Centre in 2015 by using the IXIA tool to extract normal behaviors and modern attacks, it includes normal data and 9 types of attacks with 49 features. Previous research results show that the detection of Fuzzers attacks in this dataset gives the lowest classification quality. This paper analyzes and evaluates the performance of using known ensemble techniques such as Bagging, AdaBoost, Stacking, Decorate, Random Forest and Voting to detect FUZZERS attacks on UNSW-NB15 dataset to create models. The experimental results show that the AdaBoost technique with the component classifiers using decision tree for the best classification quality with F −Measure is 96.76% compared to 94.16%, which is the best result by using single classifiers and 96.36% by using the Random Forest technique.","PeriodicalId":15444,"journal":{"name":"Journal of Computer Science and Cybernetics","volume":"24 1","pages":"173-185"},"PeriodicalIF":0.0,"publicationDate":"2020-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88260329","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2020-05-11DOI: 10.15625/1813-9663/36/2/14353
Trinh D. D. Nguyen, Loan T. T. Nguyen, Quyen Tran, Bay Vo
In business, most of companies focus on growing their profits. Besides considering profit from each product, they also focus on the relationship among products in order to support effective decision making, gain more profits and attract their customers, e.g. shelf arrangement, product displays, or product marketing, etc. Some high utility association rules have been proposed, however, they consume much memory and require long time processing. This paper proposes LHAR (Lattice-based for mining High utility Association Rules) algorithm to mine high utility association rules based on a lattice of high utility itemsets. The LHAR algorithm aims to generates high utility association rules during the process of building lattice of high utility itemsets, and thus it needs less memory and runtime
在商业中,大多数公司都专注于增加利润。除了考虑每个产品的利润,他们也关注产品之间的关系,以支持有效的决策,获得更多的利润和吸引他们的客户,如货架安排,产品展示,或产品营销等。提出了一些高实用的关联规则,但它们占用内存大,处理时间长。提出了基于高效用项集格的高效用关联规则挖掘算法LHAR (lattice -based for mining High utility Association Rules)。LHAR算法的目的是在构建高效用项集格的过程中生成高效用关联规则,从而减少了对内存和运行时间的需求
{"title":"AN EFFICIENT ALGORITHM FORMINING HIGH UTILITY ASSOCIATION RULES FROM LATTICE","authors":"Trinh D. D. Nguyen, Loan T. T. Nguyen, Quyen Tran, Bay Vo","doi":"10.15625/1813-9663/36/2/14353","DOIUrl":"https://doi.org/10.15625/1813-9663/36/2/14353","url":null,"abstract":"In business, most of companies focus on growing their profits. Besides considering profit from each product, they also focus on the relationship among products in order to support effective decision making, gain more profits and attract their customers, e.g. shelf arrangement, product displays, or product marketing, etc. Some high utility association rules have been proposed, however, they consume much memory and require long time processing. This paper proposes LHAR (Lattice-based for mining High utility Association Rules) algorithm to mine high utility association rules based on a lattice of high utility itemsets. The LHAR algorithm aims to generates high utility association rules during the process of building lattice of high utility itemsets, and thus it needs less memory and runtime","PeriodicalId":15444,"journal":{"name":"Journal of Computer Science and Cybernetics","volume":"30 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72646979","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2020-05-11DOI: 10.15625/1813-9663/36/2/14807
Chung P. Le, Kiem Nguyen Tien, Linh V. Nguyen, Tinh Nguyen, Tung Hoang
This article highlights a robust adaptive tracking backstepping control approach for a nonholonomic wheeled mobile robot (WMR) by which the bad problems of both unknown slippage and uncertainties are dealt with. The radial basis function neural network (RBFNN) in this proposed controller assists unknown smooth nonlinear dynamic functions to be approximated. Furthermore, a technical solution is also carried out to avoid actuator saturation. The validity and efficiency of this novel controller, finally, are illustrated via comparative simulation results.
{"title":"DESIGNING A ROBUST ADAPTIVE TRACKING BACKTEPPING CONTROLLER CONSIDERING ACTUATOR SATURATION FOR A WHEELED MOBILE ROBOT TO COMPENSATE UNKNOWN SLIPPAGE","authors":"Chung P. Le, Kiem Nguyen Tien, Linh V. Nguyen, Tinh Nguyen, Tung Hoang","doi":"10.15625/1813-9663/36/2/14807","DOIUrl":"https://doi.org/10.15625/1813-9663/36/2/14807","url":null,"abstract":"This article highlights a robust adaptive tracking backstepping control approach for a nonholonomic wheeled mobile robot (WMR) by which the bad problems of both unknown slippage and uncertainties are dealt with. The radial basis function neural network (RBFNN) in this proposed controller assists unknown smooth nonlinear dynamic functions to be approximated. Furthermore, a technical solution is also carried out to avoid actuator saturation. The validity and efficiency of this novel controller, finally, are illustrated via comparative simulation results.","PeriodicalId":15444,"journal":{"name":"Journal of Computer Science and Cybernetics","volume":"14 1","pages":"187-204"},"PeriodicalIF":0.0,"publicationDate":"2020-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87876258","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2020-05-11DOI: 10.15625/1813-9663/36/2/14084
Nguyen Kim Sao, Nguyen Ngoc Hoa, Pham Van At
This paper presents a new effective reversible data hiding method based on pixel-valueordering (iGePVO-K) which is improvement of a recent GePVO-K method that recently is considered as a PVO-used method having highest embedding capacity. In comparison with GePVO-K method, iGePVO-K has the following advantages. First, the embedding capacity of the new method is higher than that of GePVO-K method by using data embedding formulas reasonably and reducing the location map size. Second, for embedding data, in the new method, each pixel value is modified at most by one, while in GePVO-K method, each pixel value may be modified by two. In fact, in the GePVO-K method, the largest pixels are modified by two for embedding bits 1 and by one for bits 0. This is also true for the smallest pixels. Meanwhile, in the proposed method, the largest pixels are modified by one for embedding bits 1 and are unchanged if embedding bits 0. Therefore, the stego-image quality in proposed method is better than that in GePVO-K method. Theoretical analysis and experiment results show that the proposed method has higher embedding capacity and better stego image quality than GePVO-K method.
{"title":"AN EFFECTIVE REVERSIBLE DATA HIDING METHOD BASED ON PIXEL-VALUE-ORDERING","authors":"Nguyen Kim Sao, Nguyen Ngoc Hoa, Pham Van At","doi":"10.15625/1813-9663/36/2/14084","DOIUrl":"https://doi.org/10.15625/1813-9663/36/2/14084","url":null,"abstract":"This paper presents a new effective reversible data hiding method based on pixel-valueordering (iGePVO-K) which is improvement of a recent GePVO-K method that recently is considered as a PVO-used method having highest embedding capacity. In comparison with GePVO-K method, iGePVO-K has the following advantages. First, the embedding capacity of the new method is higher than that of GePVO-K method by using data embedding formulas reasonably and reducing the location map size. Second, for embedding data, in the new method, each pixel value is modified at most by one, while in GePVO-K method, each pixel value may be modified by two. In fact, in the GePVO-K method, the largest pixels are modified by two for embedding bits 1 and by one for bits 0. This is also true for the smallest pixels. Meanwhile, in the proposed method, the largest pixels are modified by one for embedding bits 1 and are unchanged if embedding bits 0. Therefore, the stego-image quality in proposed method is better than that in GePVO-K method. Theoretical analysis and experiment results show that the proposed method has higher embedding capacity and better stego image quality than GePVO-K method.","PeriodicalId":15444,"journal":{"name":"Journal of Computer Science and Cybernetics","volume":"35 1","pages":"139-158"},"PeriodicalIF":0.0,"publicationDate":"2020-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91337300","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2020-02-27DOI: 10.15625/1813-9663/35/4/13749
H. Anh, Cao Van Kien
This paper proposes a novel uncertain fuzzy descriptor system which is an extension from standard T-S fuzzy system. A fixed Lyapunov function-based approach is considered and controller design for this rich class of fuzzy descriptor systems is formulated as a problem of solving a set of LMIs. The design conditions for the descriptor fuzzy system are more complicated than the standard state-space-based systems. However, the descriptor fuzzy system-based approach has the advantage of possessing fewer number of matrix inequality conditions for certain special cases. Hence, it is suitable for complex systems represented in descriptor form which is often observed in nonlinear mechanical systems.
{"title":"Novel Approach of Robust Hinf Tracking Control for Uncertain Fuzzy Descriptor Systems Using Parametric Lyapunov Function","authors":"H. Anh, Cao Van Kien","doi":"10.15625/1813-9663/35/4/13749","DOIUrl":"https://doi.org/10.15625/1813-9663/35/4/13749","url":null,"abstract":"This paper proposes a novel uncertain fuzzy descriptor system which is an extension from standard T-S fuzzy system. A fixed Lyapunov function-based approach is considered and controller design for this rich class of fuzzy descriptor systems is formulated as a problem of solving a set of LMIs. The design conditions for the descriptor fuzzy system are more complicated than the standard state-space-based systems. However, the descriptor fuzzy system-based approach has the advantage of possessing fewer number of matrix inequality conditions for certain special cases. Hence, it is suitable for complex systems represented in descriptor form which is often observed in nonlinear mechanical systems.","PeriodicalId":15444,"journal":{"name":"Journal of Computer Science and Cybernetics","volume":"82 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83965686","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2020-02-27DOI: 10.15625/1813-9663/36/1/14557
Le Ngoc Truc, N. Quyen, N. P. Quang
The paper presents a complete generalized procedure based on the Euler-Lagrange equations to build the matrix form of dynamic equations, called dynamic model, for robot manipulators. In addition, a new formulation of the Coriolis/centrifugal matrix is proposed. The link linear and angular velocities are formulated explicitly. Therefore, the translational and rotational Jacobian matrices can be derived straightforward from definition, which make the calculation of the generalized inertia matrix more convenient. By using Kronecker product, a new Coriolis/centrifugal matrix formulation is set up directly in matrix-based manner and guarantees the skew symmetry property of robot dynamic equations. This important property is usually exploited for developing many control methodologies. The validation of the proposal formulation is confirmed through the symbolic solution and simulation of a typical robot manipulator.
{"title":"DYNAMIC MODEL WITH A NEW FORMULATION OF CORIOLIS/CENTRIFUGAL MATRIX FOR ROBOT MANIPULATORS","authors":"Le Ngoc Truc, N. Quyen, N. P. Quang","doi":"10.15625/1813-9663/36/1/14557","DOIUrl":"https://doi.org/10.15625/1813-9663/36/1/14557","url":null,"abstract":"The paper presents a complete generalized procedure based on the Euler-Lagrange equations to build the matrix form of dynamic equations, called dynamic model, for robot manipulators. In addition, a new formulation of the Coriolis/centrifugal matrix is proposed. The link linear and angular velocities are formulated explicitly. Therefore, the translational and rotational Jacobian matrices can be derived straightforward from definition, which make the calculation of the generalized inertia matrix more convenient. By using Kronecker product, a new Coriolis/centrifugal matrix formulation is set up directly in matrix-based manner and guarantees the skew symmetry property of robot dynamic equations. This important property is usually exploited for developing many control methodologies. The validation of the proposal formulation is confirmed through the symbolic solution and simulation of a typical robot manipulator.","PeriodicalId":15444,"journal":{"name":"Journal of Computer Science and Cybernetics","volume":"45 1","pages":"89-104"},"PeriodicalIF":0.0,"publicationDate":"2020-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78526965","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2020-02-27DOI: 10.15625/1813-9663/1/1/14368
N. Thong, Cu Nguyen Giap, T. Tuan, P. Chuan, Pham Minh Hoang, D. Dong
Multi-attributes decision-making problem in dynamic neutrosophic environment is an open and highly-interesting research area with many potential applications in real life. The concept of the dynamic interval-valued neutrosophic set and its application for the dynamic decision-making are proposed recently, however the inter-dependence among criteria or preference is not dealt with in the proposed operations to well treat inter-dependence problems. Therefore, the definitions, mathematical operations and its properties are mentioned and discussed in detail.Then, Choquet integral-based distance between dynamic inteval-valued neutrosophic sets is defined and used to develop a new decision making model based on the proposed theory. A practical application of proposed approach is constructed and tested on the data of lecturers' performance collected from Vietnam National University (VNU) to illustrate the efficiency of new proposal.
{"title":"MODELING MULTI-CRITERIA DECISION-MAKING IN DYNAMIC NEUTROSOPHIC ENVIRONMENTS BASES ON CHOQUET INTEGRAL","authors":"N. Thong, Cu Nguyen Giap, T. Tuan, P. Chuan, Pham Minh Hoang, D. Dong","doi":"10.15625/1813-9663/1/1/14368","DOIUrl":"https://doi.org/10.15625/1813-9663/1/1/14368","url":null,"abstract":"Multi-attributes decision-making problem in dynamic neutrosophic environment is an open and highly-interesting research area with many potential applications in real life. The concept of the dynamic interval-valued neutrosophic set and its application for the dynamic decision-making are proposed recently, however the inter-dependence among criteria or preference is not dealt with in the proposed operations to well treat inter-dependence problems. Therefore, the definitions, mathematical operations and its properties are mentioned and discussed in detail.Then, Choquet integral-based distance between dynamic inteval-valued neutrosophic sets is defined and used to develop a new decision making model based on the proposed theory. A practical application of proposed approach is constructed and tested on the data of lecturers' performance collected from Vietnam National University (VNU) to illustrate the efficiency of new proposal.","PeriodicalId":15444,"journal":{"name":"Journal of Computer Science and Cybernetics","volume":"68 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91168169","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2020-02-27DOI: 10.15625/1813-9663/36/1/14347
Nguyen Vu Uyen Nhi, Lê Mạnh Thạnh
The image retrieval and semantic extraction play an important role in the multimedia systems such as geographic information system, hospital information system, digital library system, etc. Therefore, the research and development of semantic-based image retrieval (SBIR) systems have become extremely important and urgent. Major recent publications are included covering different aspects of the research in this area, including building data models, low-level image feature extraction, and deriving high-level semantic features. However, there is still no general approach for semantic-based image retrieval (SBIR), due to the diversity and complexity of high-level semantics. In order to improve the retrieval accuracy of SBIR systems, our focus research is to build a data structure for finding similar images, from that retrieving its semantic. In this paper, we proposed a data structure which is a self-balanced clustering tree named C-Tree. Firstly, a method of visual semantic analysis relied on visual features and image content is proposed on C-Tree. The building of this structure is created based on a combination of methods including hierarchical clustering and partitional clustering. Secondly, we design ontology for the image dataset and create the SPARQL (SPARQL Protocol and RDF Query Language) query by extracting semantics of image. Finally, the semantic-based image retrieval on C-Tree (SBIR CT) model is created hinging on our proposal. The experimental evaluation 20,000 images of ImageCLEF dataset indicates the effectiveness of the proposed method. These results are compared with some of recently published methods on the same dataset and demonstrate that the proposed method improves the retrieval accuracy and efficiency.
图像检索和语义提取在地理信息系统、医院信息系统、数字图书馆系统等多媒体系统中起着重要的作用。因此,基于语义的图像检索(SBIR)系统的研究和开发变得极其重要和迫切。最近的主要出版物涵盖了该领域研究的不同方面,包括构建数据模型、低级图像特征提取和派生高级语义特征。然而,由于高级语义的多样性和复杂性,目前还没有一种通用的基于语义的图像检索方法。为了提高SBIR系统的检索精度,我们的研究重点是建立一个数据结构来查找相似图像,从检索其语义开始。本文提出了一种自平衡聚类树的数据结构——C-Tree。首先,在C-Tree上提出了一种基于视觉特征和图像内容的视觉语义分析方法。该结构的构建是基于分层聚类和分区聚类的组合方法创建的。其次,对图像数据集进行本体设计,并通过提取图像语义创建SPARQL (SPARQL Protocol and RDF Query Language)查询;最后,在此基础上建立了基于语义的c树图像检索(SBIR CT)模型。对ImageCLEF数据集的2万幅图像进行了实验评估,结果表明了该方法的有效性。将这些结果与最近发表的一些方法在同一数据集上的检索结果进行了比较,表明该方法提高了检索精度和效率。
{"title":"A SELF-BALANCED CLUSTERING TREE FOR SEMANTIC-BASED IMAGE RETRIEVAL","authors":"Nguyen Vu Uyen Nhi, Lê Mạnh Thạnh","doi":"10.15625/1813-9663/36/1/14347","DOIUrl":"https://doi.org/10.15625/1813-9663/36/1/14347","url":null,"abstract":"The image retrieval and semantic extraction play an important role in the multimedia systems such as geographic information system, hospital information system, digital library system, etc. Therefore, the research and development of semantic-based image retrieval (SBIR) systems have become extremely important and urgent. Major recent publications are included covering different aspects of the research in this area, including building data models, low-level image feature extraction, and deriving high-level semantic features. However, there is still no general approach for semantic-based image retrieval (SBIR), due to the diversity and complexity of high-level semantics. In order to improve the retrieval accuracy of SBIR systems, our focus research is to build a data structure for finding similar images, from that retrieving its semantic. In this paper, we proposed a data structure which is a self-balanced clustering tree named C-Tree. Firstly, a method of visual semantic analysis relied on visual features and image content is proposed on C-Tree. The building of this structure is created based on a combination of methods including hierarchical clustering and partitional clustering. Secondly, we design ontology for the image dataset and create the SPARQL (SPARQL Protocol and RDF Query Language) query by extracting semantics of image. Finally, the semantic-based image retrieval on C-Tree (SBIR CT) model is created hinging on our proposal. The experimental evaluation 20,000 images of ImageCLEF dataset indicates the effectiveness of the proposed method. These results are compared with some of recently published methods on the same dataset and demonstrate that the proposed method improves the retrieval accuracy and efficiency.","PeriodicalId":15444,"journal":{"name":"Journal of Computer Science and Cybernetics","volume":"104 1","pages":"49-67"},"PeriodicalIF":0.0,"publicationDate":"2020-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74957678","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}