首页 > 最新文献

2010 Second International Conference on Machine Learning and Computing最新文献

英文 中文
A Survey of Semantic Similarity Methods for Ontology Based Information Retrieval 基于本体的信息检索语义相似度方法综述
Pub Date : 2010-02-09 DOI: 10.1109/ICMLC.2010.63
K. Saruladha, G. Aghila, S. Raj
This paper discusses the various approaches used for identifying semantically similar concepts in an ontology. The purpose of this survey is to explore how these similarity computation methods could assist in ontology based query expansion. This query expansion method based on the similarity function is expected to improve the retrieval effectiveness of the ontology based Information retrieval models. Various similarity computation methods fall under three categories: Edge counting, information content and node based counting. The limitations of each of these approaches have been discussed in this paper.
本文讨论了用于识别本体中语义相似概念的各种方法。本调查的目的是探索这些相似度计算方法如何帮助基于本体的查询扩展。这种基于相似度函数的查询扩展方法有望提高基于本体的信息检索模型的检索效率。各种相似度计算方法分为三类:边缘计数、信息内容计数和基于节点计数。本文讨论了每种方法的局限性。
{"title":"A Survey of Semantic Similarity Methods for Ontology Based Information Retrieval","authors":"K. Saruladha, G. Aghila, S. Raj","doi":"10.1109/ICMLC.2010.63","DOIUrl":"https://doi.org/10.1109/ICMLC.2010.63","url":null,"abstract":"This paper discusses the various approaches used for identifying semantically similar concepts in an ontology. The purpose of this survey is to explore how these similarity computation methods could assist in ontology based query expansion. This query expansion method based on the similarity function is expected to improve the retrieval effectiveness of the ontology based Information retrieval models. Various similarity computation methods fall under three categories: Edge counting, information content and node based counting. The limitations of each of these approaches have been discussed in this paper.","PeriodicalId":423912,"journal":{"name":"2010 Second International Conference on Machine Learning and Computing","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122006603","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 51
The Design and Implementation of a Practical Meta-Heuristic for the Detection and Identification of Denial-of-Service Attack Using Hybrid Approach 一种基于混合方法的实用元启发式拒绝服务攻击检测与识别方法的设计与实现
Pub Date : 2010-02-09 DOI: 10.1109/ICMLC.2010.46
Hsia-Hsiang Chen, Wuu Yang
Network attacks are occurring continuously day after day. The researchers are expected to find the solution by identifying the address of source. We propose the IP traceback ant colony system (ITACS) algorithm to solve the IP traceback of denial of service (DoS) problem. The ITACS is novel attempted to apply in solving the problem. It is a meta- heuristic algorithm, which is a technique applies so that attack detection and attack identification can be implemented. The proposed algorithm has improved by the previous one to conquer this problem successfully. We obtained the data set of topology from one of famous research organizations for the experiment. The parameters of algorithm are considered by packet contents in topology. In the meanwhile, we discussed the increment of traffic condition. By the experiment, the examples of increment of traffic are above average 70%. The results show that the performance of ITACS algorithm is efficient and accurate. Furthermore, the proposed algorithm has also nature of robust for the problem. Future work may even be extended to study the other behaviors of organisms from derivations of meta-heuristic algorithm.
网络攻击日益频繁。研究人员希望通过确定源地址来找到解决方案。针对IP溯源拒绝服务(DoS)问题,提出了IP溯源蚁群系统(ITACS)算法。ITACS是解决这一问题的新颖尝试。它是一种元启发式算法,是一种用于实现攻击检测和攻击识别的技术。该算法在原有算法的基础上进行了改进,成功地解决了这一问题。我们从一个著名的研究机构获得了拓扑数据集用于实验。算法的参数由拓扑中的数据包内容来考虑。同时,对交通条件的增量进行了讨论。通过实验,流量增量的实例均在70%以上。结果表明,ITACS算法具有高效、准确的性能。此外,该算法对该问题还具有鲁棒性。未来的工作甚至可以扩展到从元启发式算法的衍生研究生物体的其他行为。
{"title":"The Design and Implementation of a Practical Meta-Heuristic for the Detection and Identification of Denial-of-Service Attack Using Hybrid Approach","authors":"Hsia-Hsiang Chen, Wuu Yang","doi":"10.1109/ICMLC.2010.46","DOIUrl":"https://doi.org/10.1109/ICMLC.2010.46","url":null,"abstract":"Network attacks are occurring continuously day after day. The researchers are expected to find the solution by identifying the address of source. We propose the IP traceback ant colony system (ITACS) algorithm to solve the IP traceback of denial of service (DoS) problem. The ITACS is novel attempted to apply in solving the problem. It is a meta- heuristic algorithm, which is a technique applies so that attack detection and attack identification can be implemented. The proposed algorithm has improved by the previous one to conquer this problem successfully. We obtained the data set of topology from one of famous research organizations for the experiment. The parameters of algorithm are considered by packet contents in topology. In the meanwhile, we discussed the increment of traffic condition. By the experiment, the examples of increment of traffic are above average 70%. The results show that the performance of ITACS algorithm is efficient and accurate. Furthermore, the proposed algorithm has also nature of robust for the problem. Future work may even be extended to study the other behaviors of organisms from derivations of meta-heuristic algorithm.","PeriodicalId":423912,"journal":{"name":"2010 Second International Conference on Machine Learning and Computing","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126012812","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Genetically Improved PSO Algorithm for Efficient Data Clustering 高效数据聚类的遗传改进粒子群算法
Pub Date : 2010-02-09 DOI: 10.1109/ICMLC.2010.19
Rehab F. Abdel-Kader
Clustering is an important research topic in data mining that appears in a wide range of unsupervised classification applications. Partitional clustering algorithms such as the k-means algorithm are the most popular for clustering large datasets. The major problem with the k-means algorithm is that it is sensitive to the selection of the initial partitions and it may converge to local optima. In this paper, we present a hybrid two-phase GAI-PSO+k-means data clustering algorithm that performs fast data clustering and can avoid premature convergence to local optima. In the first phase we utilize the new genetically improved particle swarm optimization algorithm (GAI-PSO) which is a population-based heuristic search technique modeled on the hybrid of cultural and social rules derived from the analysis of the swarm intelligence (PSO) and the concepts of natural selection and evolution (GA). The GAI-PSO combines the standard velocity and position update rules of PSOs with the ideas of selection, mutation and crossover from GAs. The GAI-PSO algorithm searches the solution space to find the optimal initial cluster centroids for the next phase. The second phase is a local refining stage utilizing the k-means algorithm which can efficiently converge to the optimal solution. The proposed algorithm combines the ability of the globalized searching of the evolutionary algorithms and the fast convergence of the k-means algorithm and can avoid the drawback of both. The performance of the proposed algorithm is evaluated through several benchmark datasets. The experimental results show that the proposed algorithm is highly forceful and outperforms the previous approaches such as SA, ACO, PSO and k-means for the partitional clustering problem.
聚类是数据挖掘领域的一个重要研究课题,广泛应用于无监督分类领域。分割聚类算法,如k-means算法是聚类大型数据集最流行的算法。k-means算法的主要问题是它对初始分区的选择很敏感,并且可能收敛到局部最优。本文提出了一种混合的两阶段GAI-PSO+k-means数据聚类算法,该算法具有快速聚类和避免过早收敛到局部最优的优点。在第一阶段,我们使用了新的遗传改进粒子群优化算法(遗传改进粒子群优化算法),这是一种基于群体的启发式搜索技术,它基于对群体智能(PSO)和自然选择与进化(GA)概念的分析而得出的文化和社会规则的混合模型。该算法将粒子群的标准速度和位置更新规则与遗传算法的选择、变异和交叉思想相结合。该算法通过搜索解空间来寻找下一阶段的最优初始聚类质心。第二阶段是局部细化阶段,利用k-means算法可以有效收敛到最优解。该算法结合了进化算法的全球化搜索能力和k-means算法的快速收敛性,避免了两者的缺点。通过多个基准数据集对该算法的性能进行了评估。实验结果表明,该算法具有较强的求解力,在分割聚类问题上优于SA、ACO、PSO和k-means等方法。
{"title":"Genetically Improved PSO Algorithm for Efficient Data Clustering","authors":"Rehab F. Abdel-Kader","doi":"10.1109/ICMLC.2010.19","DOIUrl":"https://doi.org/10.1109/ICMLC.2010.19","url":null,"abstract":"Clustering is an important research topic in data mining that appears in a wide range of unsupervised classification applications. Partitional clustering algorithms such as the k-means algorithm are the most popular for clustering large datasets. The major problem with the k-means algorithm is that it is sensitive to the selection of the initial partitions and it may converge to local optima. In this paper, we present a hybrid two-phase GAI-PSO+k-means data clustering algorithm that performs fast data clustering and can avoid premature convergence to local optima. In the first phase we utilize the new genetically improved particle swarm optimization algorithm (GAI-PSO) which is a population-based heuristic search technique modeled on the hybrid of cultural and social rules derived from the analysis of the swarm intelligence (PSO) and the concepts of natural selection and evolution (GA). The GAI-PSO combines the standard velocity and position update rules of PSOs with the ideas of selection, mutation and crossover from GAs. The GAI-PSO algorithm searches the solution space to find the optimal initial cluster centroids for the next phase. The second phase is a local refining stage utilizing the k-means algorithm which can efficiently converge to the optimal solution. The proposed algorithm combines the ability of the globalized searching of the evolutionary algorithms and the fast convergence of the k-means algorithm and can avoid the drawback of both. The performance of the proposed algorithm is evaluated through several benchmark datasets. The experimental results show that the proposed algorithm is highly forceful and outperforms the previous approaches such as SA, ACO, PSO and k-means for the partitional clustering problem.","PeriodicalId":423912,"journal":{"name":"2010 Second International Conference on Machine Learning and Computing","volume":"87 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116570389","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 67
Adapting Moments for Handwritten Kannada Kagunita Recognition 手写体Kannada Kagunita识别的自适应时刻
Pub Date : 2010-02-09 DOI: 10.1109/ICMLC.2010.51
L. Ragha, M. Sasikumar
The Handwriting character recognition (HCR) for Indian Languages is an important problem where there is relatively little work has been done. In this paper, we investigate the use of moments features on Kannada Kagunita. Kannada characters are curved in nature with some kind of symmetric structure observed in the shape. This information can be best extracted as a feature if we extract moment features from the directional images. To recognize a Kagunita, we need to identify the vowel and the consonant present in the image. So we are finding 4 directional images using Gabor wavelets from the dynamically preprocessed original image. We analyze the Kagunita set and identify the regions with vowel information and consonant information and cut these portions from the preprocessed original image and form a set of cut images. We then extract moments features from them. These features are trained and tested for both vowel and Kagunita recognition on Multi Layer Perceptron with Back Propagation Neural Network. The recognition results for vowels is average 85% and consonants is 59% when tested on separate test data with moments features from directional images and cut images.
印度语言的手写字符识别(HCR)是一个重要的问题,目前已经完成的工作相对较少。在本文中,我们研究了弯矩特征在kanadada Kagunita中的应用。卡纳达文字本质上是弯曲的,在形状上可以观察到某种对称结构。如果我们从方向图像中提取矩特征,可以最好地将这些信息作为特征提取出来。为了识别一个Kagunita,我们需要识别图像中的元音和辅音。因此,我们使用Gabor小波从动态预处理的原始图像中找到4个方向图像。我们对Kagunita集合进行分析,识别出含有元音信息和辅音信息的区域,并将这些区域从预处理后的原始图像中剪切出来,形成一组剪切图像。然后从中提取矩特征。利用反向传播神经网络对这些特征进行训练和测试,并在多层感知器上进行元音和元音识别。在具有方向图像和剪切图像矩特征的单独测试数据上进行测试,元音的平均识别率为85%,辅音的平均识别率为59%。
{"title":"Adapting Moments for Handwritten Kannada Kagunita Recognition","authors":"L. Ragha, M. Sasikumar","doi":"10.1109/ICMLC.2010.51","DOIUrl":"https://doi.org/10.1109/ICMLC.2010.51","url":null,"abstract":"The Handwriting character recognition (HCR) for Indian Languages is an important problem where there is relatively little work has been done. In this paper, we investigate the use of moments features on Kannada Kagunita. Kannada characters are curved in nature with some kind of symmetric structure observed in the shape. This information can be best extracted as a feature if we extract moment features from the directional images. To recognize a Kagunita, we need to identify the vowel and the consonant present in the image. So we are finding 4 directional images using Gabor wavelets from the dynamically preprocessed original image. We analyze the Kagunita set and identify the regions with vowel information and consonant information and cut these portions from the preprocessed original image and form a set of cut images. We then extract moments features from them. These features are trained and tested for both vowel and Kagunita recognition on Multi Layer Perceptron with Back Propagation Neural Network. The recognition results for vowels is average 85% and consonants is 59% when tested on separate test data with moments features from directional images and cut images.","PeriodicalId":423912,"journal":{"name":"2010 Second International Conference on Machine Learning and Computing","volume":"64 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129924064","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Application of Lagrangian Twin Support Vector Machines for Classification 拉格朗日孪生支持向量机在分类中的应用
Pub Date : 2010-02-09 DOI: 10.1109/ICMLC.2010.40
S. Balasundaram, N. Kapil
In this paper a new iterative approach is proposed for solving the Lagrangian formulation of twin support vector machine classifiers. The main advantage of our method is that rather than solving a quadratic programming problem as in the case of the standard support vector machine the inverse of a matrix of size equals to the number of input examples needs to be determined at the very beginning of the algorithm. The convergence of the algorithm is stated. Experiments have been performed on a number of interesting datasets. The predicted results are in good agreement with the observed values clearly demonstrates the applicability of the proposed method.
本文提出了一种新的迭代方法来求解双支持向量机分类器的拉格朗日公式。我们的方法的主要优点是,不像标准支持向量机那样解决二次规划问题,需要在算法的一开始就确定大小等于输入示例数量的矩阵的逆。说明了算法的收敛性。在许多有趣的数据集上进行了实验。预测结果与实测值吻合较好,表明了所提方法的适用性。
{"title":"Application of Lagrangian Twin Support Vector Machines for Classification","authors":"S. Balasundaram, N. Kapil","doi":"10.1109/ICMLC.2010.40","DOIUrl":"https://doi.org/10.1109/ICMLC.2010.40","url":null,"abstract":"In this paper a new iterative approach is proposed for solving the Lagrangian formulation of twin support vector machine classifiers. The main advantage of our method is that rather than solving a quadratic programming problem as in the case of the standard support vector machine the inverse of a matrix of size equals to the number of input examples needs to be determined at the very beginning of the algorithm. The convergence of the algorithm is stated. Experiments have been performed on a number of interesting datasets. The predicted results are in good agreement with the observed values clearly demonstrates the applicability of the proposed method.","PeriodicalId":423912,"journal":{"name":"2010 Second International Conference on Machine Learning and Computing","volume":"64 9","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120969703","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
A Novel Data Generation Approach for Digital Forensic Application in Data Mining 一种新的数字取证数据生成方法在数据挖掘中的应用
Pub Date : 2010-02-09 DOI: 10.1109/ICMLC.2010.24
Veena H. Bhat, Prasanth G. Rao, Abhilash R.V., P. D. Shenoy, Venugopal K.R., L. Patnaik
With the rapid advancements in information and communication technology in the world, crimes committed are also becoming technically intensive. When crimes committed use digital devices, forensic examiners have to adopt practical frameworks and methods for recovering data for analysis as evidence. Data Generation, Data Warehousing and Data Mining, are the three essential features involved in this process. This paper proposes a unique way of generating, storing and analyzing data, retrieved from digital devices which pose as evidence in forensic analysis. A statistical approach is used in validating the reliability of the pre-processed data. This work proposes a practical framework for digital forensics on flash drives.
随着世界信息通信技术的飞速发展,犯罪的技术含量也越来越高。当犯罪使用数字设备时,法医必须采用实用的框架和方法来恢复数据以作为证据进行分析。数据生成、数据仓库和数据挖掘是这一过程中涉及的三个基本特征。本文提出了一种独特的方法来生成、存储和分析从数字设备中检索的数据,这些数据在法医分析中作为证据。采用统计方法验证预处理数据的可靠性。这项工作为闪存驱动器上的数字取证提出了一个实用的框架。
{"title":"A Novel Data Generation Approach for Digital Forensic Application in Data Mining","authors":"Veena H. Bhat, Prasanth G. Rao, Abhilash R.V., P. D. Shenoy, Venugopal K.R., L. Patnaik","doi":"10.1109/ICMLC.2010.24","DOIUrl":"https://doi.org/10.1109/ICMLC.2010.24","url":null,"abstract":"With the rapid advancements in information and communication technology in the world, crimes committed are also becoming technically intensive. When crimes committed use digital devices, forensic examiners have to adopt practical frameworks and methods for recovering data for analysis as evidence. Data Generation, Data Warehousing and Data Mining, are the three essential features involved in this process. This paper proposes a unique way of generating, storing and analyzing data, retrieved from digital devices which pose as evidence in forensic analysis. A statistical approach is used in validating the reliability of the pre-processed data. This work proposes a practical framework for digital forensics on flash drives.","PeriodicalId":423912,"journal":{"name":"2010 Second International Conference on Machine Learning and Computing","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127643451","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Traffic Modeling with Multi Agent Bayesian and Causal Networks and Performance Prediction for Changed Setting System 基于多智能体贝叶斯和因果网络的交通建模及变设置系统的性能预测
Pub Date : 2010-02-09 DOI: 10.1109/ICMLC.2010.34
R. Maarefdoust, S. Rahati
The traffic modeling is one of the effective methods of detecting and evaluating the urban traffic. The effect of uncertain factors such as the different behavior of a human society would count as an intricacy of the issue and would cause some problems for modeling. Level crossroads are one of the important sections in an urban traffic control system and are usually controlled by traffic lights. In this study, an attempt has been made to model the traffic of an important crossroads in Mashhad city using intelligent elements in a multi-agent environment and a large amount of real data. For this purpose, the total traffic behavior at the intersection was first modeled based on the Bayesian networks structures. Then, effective factors have been modeled using the probabilistic causal networks. Results of the evaluation of the model show that this model is able to measure system efficiency according to variances in the crossroads adjustments. Also, this model is cheaper and less time-consuming. On this basis, this modeling can be used for the evaluating and even predicting the efficacy of the traffic control system in the crossroads. The data used in this study have been collected by the SCATS software in Mashhad Traffic Control Center. The Weka software has been used for training and evaluations with the Bayesian and causal probabilistic networks.
交通建模是检测和评价城市交通的有效方法之一。不确定因素的影响,如人类社会的不同行为,将被视为问题的复杂性,并会给建模带来一些问题。水平十字路口是城市交通控制系统中的重要路段之一,通常由交通信号灯进行控制。本研究尝试利用多智能体环境下的智能元素和大量真实数据,对马什哈德市重要十字路口的交通进行建模。为此,首先建立了基于贝叶斯网络结构的交叉口总交通行为模型。然后,利用概率因果网络对有效因素进行建模。对模型的评价结果表明,该模型能够根据十字路口调整的方差来衡量系统效率。此外,这种模式更便宜,更省时。在此基础上,该模型可用于评价甚至预测十字路口交通控制系统的有效性。本研究使用的数据是由马什哈德交通控制中心的SCATS软件收集的。Weka软件已被用于贝叶斯和因果概率网络的训练和评估。
{"title":"Traffic Modeling with Multi Agent Bayesian and Causal Networks and Performance Prediction for Changed Setting System","authors":"R. Maarefdoust, S. Rahati","doi":"10.1109/ICMLC.2010.34","DOIUrl":"https://doi.org/10.1109/ICMLC.2010.34","url":null,"abstract":"The traffic modeling is one of the effective methods of detecting and evaluating the urban traffic. The effect of uncertain factors such as the different behavior of a human society would count as an intricacy of the issue and would cause some problems for modeling. Level crossroads are one of the important sections in an urban traffic control system and are usually controlled by traffic lights. In this study, an attempt has been made to model the traffic of an important crossroads in Mashhad city using intelligent elements in a multi-agent environment and a large amount of real data. For this purpose, the total traffic behavior at the intersection was first modeled based on the Bayesian networks structures. Then, effective factors have been modeled using the probabilistic causal networks. Results of the evaluation of the model show that this model is able to measure system efficiency according to variances in the crossroads adjustments. Also, this model is cheaper and less time-consuming. On this basis, this modeling can be used for the evaluating and even predicting the efficacy of the traffic control system in the crossroads. The data used in this study have been collected by the SCATS software in Mashhad Traffic Control Center. The Weka software has been used for training and evaluations with the Bayesian and causal probabilistic networks.","PeriodicalId":423912,"journal":{"name":"2010 Second International Conference on Machine Learning and Computing","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124786127","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
SGA Implementation Using Integer Arrays for Storage of Binary Strings 使用整数数组存储二进制字符串的SGA实现
Pub Date : 2010-02-09 DOI: 10.1109/ICMLC.2010.62
P. Kanchan, Rio G. L. D'Souza
The Simple Genetic Algorithm evaluates a group of binary strings on the basis of their fitness, performs crossover and mutation on them and tries to generate a group having maximum fitness. The usual method used for implementing the SGA is by using character arrays for storage of binary strings. But, this method has some disadvantages. The SGA implementation can be termed a success if the average fitness of the new generation is more than the initial average fitness. In this paper, we plan to implement the SGA using integer arrays for storage of binary strings. Then, we plan to compare the initial average fitness with the final average fitness so that the working of SGA can be verified. We have written the application such that varying population sizes can be given to check the correctness of the SGA algorithm.
简单遗传算法根据一组二进制字符串的适应度对其进行评估,并对其进行交叉和突变,以产生适应度最大的一组。实现SGA的常用方法是使用字符数组来存储二进制字符串。但是,这种方法有一些缺点。如果新一代的平均适应度大于初始平均适应度,则SGA实现可以称为成功。在本文中,我们计划使用整数数组来存储二进制字符串来实现SGA。然后,我们计划将初始平均适应度与最终平均适应度进行比较,以验证SGA的有效性。我们已经编写了应用程序,以便可以给出不同的人口大小来检查SGA算法的正确性。
{"title":"SGA Implementation Using Integer Arrays for Storage of Binary Strings","authors":"P. Kanchan, Rio G. L. D'Souza","doi":"10.1109/ICMLC.2010.62","DOIUrl":"https://doi.org/10.1109/ICMLC.2010.62","url":null,"abstract":"The Simple Genetic Algorithm evaluates a group of binary strings on the basis of their fitness, performs crossover and mutation on them and tries to generate a group having maximum fitness. The usual method used for implementing the SGA is by using character arrays for storage of binary strings. But, this method has some disadvantages. The SGA implementation can be termed a success if the average fitness of the new generation is more than the initial average fitness. In this paper, we plan to implement the SGA using integer arrays for storage of binary strings. Then, we plan to compare the initial average fitness with the final average fitness so that the working of SGA can be verified. We have written the application such that varying population sizes can be given to check the correctness of the SGA algorithm.","PeriodicalId":423912,"journal":{"name":"2010 Second International Conference on Machine Learning and Computing","volume":"98 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121240876","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Content-Based Classification and Retrieval of Wild Animal Sounds Using Feature Selection Algorithm 基于内容的野生动物声音特征选择分类与检索
Pub Date : 2010-02-09 DOI: 10.1109/ICMLC.2010.11
S. Gunasekaran, K. Revathy
Automatic animal sound classification and retrieval is very helpful for bioacoustic and audio retrieval applications. In this paper we propose a system to define and extract a set of acoustic features from all archived wild animal sound recordings that is used in subsequent feature selection, classification and retrieval tasks. The database consisted of sounds of six wild animals. The Fractal Dimension analysis based segmentation was selected due to its ability to select the right portion of signal for extracting the features. The feature vectors of the proposed algorithm consist of spectral, temporal and perceptual features of the animal vocalizations. The minimal Redundancy, Maximal Relevance (mRMR) feature selection analysis was exploited to increase the classification accuracy at a compact set of features. These features were used as the inputs of two neural networks, the k-Nearest Neighbor (kNN), the Multi-Layer Perceptron (MLP) and its fusion. The proposed system provides quite robust approach for classification and retrieval purposes, especially for the wild animal sounds.
动物声音的自动分类和检索对生物声学和音频检索应用有很大的帮助。在本文中,我们提出了一个系统来定义和提取一组声音特征从所有存档的野生动物录音,用于随后的特征选择,分类和检索任务。这个数据库由六种野生动物的声音组成。选择基于分形维数分析的分割方法,因为它能够选择合适的信号部分进行特征提取。该算法的特征向量由动物发声的光谱特征、时间特征和感知特征组成。利用最小冗余,最大相关性(mRMR)特征选择分析来提高紧凑特征集的分类精度。这些特征被用作两个神经网络的输入,即k-最近邻(kNN)、多层感知器(MLP)及其融合。该系统为分类和检索提供了可靠的方法,特别是对野生动物的声音。
{"title":"Content-Based Classification and Retrieval of Wild Animal Sounds Using Feature Selection Algorithm","authors":"S. Gunasekaran, K. Revathy","doi":"10.1109/ICMLC.2010.11","DOIUrl":"https://doi.org/10.1109/ICMLC.2010.11","url":null,"abstract":"Automatic animal sound classification and retrieval is very helpful for bioacoustic and audio retrieval applications. In this paper we propose a system to define and extract a set of acoustic features from all archived wild animal sound recordings that is used in subsequent feature selection, classification and retrieval tasks. The database consisted of sounds of six wild animals. The Fractal Dimension analysis based segmentation was selected due to its ability to select the right portion of signal for extracting the features. The feature vectors of the proposed algorithm consist of spectral, temporal and perceptual features of the animal vocalizations. The minimal Redundancy, Maximal Relevance (mRMR) feature selection analysis was exploited to increase the classification accuracy at a compact set of features. These features were used as the inputs of two neural networks, the k-Nearest Neighbor (kNN), the Multi-Layer Perceptron (MLP) and its fusion. The proposed system provides quite robust approach for classification and retrieval purposes, especially for the wild animal sounds.","PeriodicalId":423912,"journal":{"name":"2010 Second International Conference on Machine Learning and Computing","volume":"2004 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127332554","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
SVM-Based Cost-sensitive Classification Algorithm with Error Cost and Class-dependent Reject Cost 基于svm的错误代价和类相关拒绝代价的代价敏感分类算法
Pub Date : 2010-02-09 DOI: 10.1109/ICMLC.2010.27
Enhui Zheng, Chao Zou, Jian Sun, Le Chen, Ping Li
In such real data mining applications as medical diagnosis, fraud detection and fault classification, and so on, the two problems that the error cost is expensive and the reject cost is class-dependent are often encountered. In order to overcome those problems, firstly, the general mathematical description of the Binary Classification Problem with Error Cost and Class-dependent Reject Cost (BCP-EC2RC) is proposed. Secondly, as one of implementation methods of BCP-EC2RC, the new algorithm, named as Cost-sensitive Support Vector Machines with the Error Cost and the Class-dependent Reject Cost (CSVM-EC2RC), is presented. The CSVM-EC2RC algorithm involves two stages: estimating the classification reliability based on trained SVM classifier, and determining the optimal reject rate of positive class and negative class by minimizing the average cost based on the given error cost and class-dependent reject cost. The experiment studies based on a benchmark data set illustrate that the proposed algorithm is effective.
在医疗诊断、欺诈检测和故障分类等实际数据挖掘应用中,经常会遇到错误代价昂贵和拒绝代价依赖于类别的两个问题。为了克服这些问题,首先提出了带有错误代价和类相关拒绝代价的二元分类问题(BCP-EC2RC)的一般数学描述。其次,作为BCP-EC2RC的一种实现方法,提出了带有错误代价和类相关拒绝代价的代价敏感支持向量机(CSVM-EC2RC)算法。CSVM-EC2RC算法包括两个阶段:基于训练好的SVM分类器估计分类可靠性;基于给定的错误代价和类相关的拒绝代价,通过最小化平均代价来确定正类和负类的最优拒绝率。基于一个基准数据集的实验研究表明,该算法是有效的。
{"title":"SVM-Based Cost-sensitive Classification Algorithm with Error Cost and Class-dependent Reject Cost","authors":"Enhui Zheng, Chao Zou, Jian Sun, Le Chen, Ping Li","doi":"10.1109/ICMLC.2010.27","DOIUrl":"https://doi.org/10.1109/ICMLC.2010.27","url":null,"abstract":"In such real data mining applications as medical diagnosis, fraud detection and fault classification, and so on, the two problems that the error cost is expensive and the reject cost is class-dependent are often encountered. In order to overcome those problems, firstly, the general mathematical description of the Binary Classification Problem with Error Cost and Class-dependent Reject Cost (BCP-EC2RC) is proposed. Secondly, as one of implementation methods of BCP-EC2RC, the new algorithm, named as Cost-sensitive Support Vector Machines with the Error Cost and the Class-dependent Reject Cost (CSVM-EC2RC), is presented. The CSVM-EC2RC algorithm involves two stages: estimating the classification reliability based on trained SVM classifier, and determining the optimal reject rate of positive class and negative class by minimizing the average cost based on the given error cost and class-dependent reject cost. The experiment studies based on a benchmark data set illustrate that the proposed algorithm is effective.","PeriodicalId":423912,"journal":{"name":"2010 Second International Conference on Machine Learning and Computing","volume":"123 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133048615","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
2010 Second International Conference on Machine Learning and Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1