首页 > 最新文献

2017 Tenth International Conference on Contemporary Computing (IC3)最新文献

英文 中文
Clustering based minimum spanning tree algorithm 基于聚类的最小生成树算法
Pub Date : 2017-08-01 DOI: 10.1109/IC3.2017.8284349
Sakshi Saxena, Priyanka Verma, D. Rajpoot
Data mining is a technique used to process information from a big dataset and converting it into a reasonable form for supplementary use. Clustering is a mining technique used in data mining. The goal of clustering is to discover the groupings of a set of points, patterns or objects. Minimum Spanning Tree (MST) based clustering algorithms are successfully used for detecting clusters. In this paper we have focused on minimizing the time complexity for constructing MST by using clustering. The proposed algorithm tries to minimize the time complexity by constructing a MST in two stages. In divide stage, the given dataset is divided in various clusters. In the conquer stage, for every cluster, local MSTs are created and then these MSTs are combined to obtain the final MST by using Midpoint MST algorithm. Experimental results show that the proposed MST algorithm is computationally efficient.
数据挖掘是一种用于处理大数据集中的信息并将其转换为合理形式以供补充使用的技术。聚类是一种用于数据挖掘的挖掘技术。聚类的目标是发现一组点、模式或对象的分组。基于最小生成树(MST)的聚类算法成功地用于聚类检测。本文主要研究如何利用聚类方法最小化构造MST的时间复杂度。该算法通过分两个阶段构造MST来最小化时间复杂度。在划分阶段,将给定的数据集划分为不同的聚类。在征服阶段,对每个集群先创建局部MST,然后使用Midpoint MST算法将这些MST组合得到最终的MST。实验结果表明,该算法具有较高的计算效率。
{"title":"Clustering based minimum spanning tree algorithm","authors":"Sakshi Saxena, Priyanka Verma, D. Rajpoot","doi":"10.1109/IC3.2017.8284349","DOIUrl":"https://doi.org/10.1109/IC3.2017.8284349","url":null,"abstract":"Data mining is a technique used to process information from a big dataset and converting it into a reasonable form for supplementary use. Clustering is a mining technique used in data mining. The goal of clustering is to discover the groupings of a set of points, patterns or objects. Minimum Spanning Tree (MST) based clustering algorithms are successfully used for detecting clusters. In this paper we have focused on minimizing the time complexity for constructing MST by using clustering. The proposed algorithm tries to minimize the time complexity by constructing a MST in two stages. In divide stage, the given dataset is divided in various clusters. In the conquer stage, for every cluster, local MSTs are created and then these MSTs are combined to obtain the final MST by using Midpoint MST algorithm. Experimental results show that the proposed MST algorithm is computationally efficient.","PeriodicalId":147099,"journal":{"name":"2017 Tenth International Conference on Contemporary Computing (IC3)","volume":"78 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114383254","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Currency recognition system using image processing 货币识别系统采用图像处理
Pub Date : 2017-08-01 DOI: 10.1109/IC3.2017.8284300
Vedasamhitha Abburu, Saumya Gupta, S. R. Rimitha, Manjunath Mulimani, S. Koolagudi
In this paper, we propose a system for automated currency recognition using image processing techniques. The proposed method can be used for recognizing both the country or origin as well as the denomination or value of a given banknote. Only paper currencies have been considered. This method works by first identifying the country of origin using certain predefined areas of interest, and then extracting the denomination value using characteristics such as size, color, or text on the note, depending on how much the notes within the same country differ. We have considered 20 of the most traded currencies, as well as their denominations. Our system is able to accurately and quickly identify test notes.
在本文中,我们提出了一个使用图像处理技术的自动货币识别系统。所提出的方法既可用于识别国家或原产,也可用于识别给定钞票的面额或价值。只考虑了纸币。这种方法的工作原理是,首先使用某些预定义的兴趣区域识别原产国,然后根据同一国家的纸币差异的大小、颜色或文字等特征提取面值值。我们考虑了20种交易最频繁的货币及其面额。我们的系统能够准确、快速地识别测试笔记。
{"title":"Currency recognition system using image processing","authors":"Vedasamhitha Abburu, Saumya Gupta, S. R. Rimitha, Manjunath Mulimani, S. Koolagudi","doi":"10.1109/IC3.2017.8284300","DOIUrl":"https://doi.org/10.1109/IC3.2017.8284300","url":null,"abstract":"In this paper, we propose a system for automated currency recognition using image processing techniques. The proposed method can be used for recognizing both the country or origin as well as the denomination or value of a given banknote. Only paper currencies have been considered. This method works by first identifying the country of origin using certain predefined areas of interest, and then extracting the denomination value using characteristics such as size, color, or text on the note, depending on how much the notes within the same country differ. We have considered 20 of the most traded currencies, as well as their denominations. Our system is able to accurately and quickly identify test notes.","PeriodicalId":147099,"journal":{"name":"2017 Tenth International Conference on Contemporary Computing (IC3)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117024880","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
Algorithms for projecting a bipartite network 投影二部网络的算法
Pub Date : 2017-08-01 DOI: 10.1109/IC3.2017.8284345
Suman Banerjee, M. Jenamani, D. K. Pratihar
Bipartite Graph or bipartite network (also known as two mode network) is often a general model of many real life complex networks and systems. However, due to the lack of analysis techniques, it is often converted into a unipartite network by one mode projection. Thus, performing faster one mode projection will lead to the faster analysis of input bipartite graph. In this paper, we have taken up the problem of one mode projection and presented three algorithms. All the algorithms have different working principles. The proposed algorithms have also been implemented on three benchmark datasets and execution times are reported.
二部图或二部网络(也称为双模网络)通常是许多现实生活中复杂网络和系统的一般模型。然而,由于缺乏分析技术,通常通过单模投影将其转换为单部网络。因此,执行更快的一模投影将导致更快的分析输入二部图。本文讨论了单模投影问题,并给出了三种算法。所有的算法都有不同的工作原理。所提出的算法也在三个基准数据集上实现,并报告了执行时间。
{"title":"Algorithms for projecting a bipartite network","authors":"Suman Banerjee, M. Jenamani, D. K. Pratihar","doi":"10.1109/IC3.2017.8284345","DOIUrl":"https://doi.org/10.1109/IC3.2017.8284345","url":null,"abstract":"Bipartite Graph or bipartite network (also known as two mode network) is often a general model of many real life complex networks and systems. However, due to the lack of analysis techniques, it is often converted into a unipartite network by one mode projection. Thus, performing faster one mode projection will lead to the faster analysis of input bipartite graph. In this paper, we have taken up the problem of one mode projection and presented three algorithms. All the algorithms have different working principles. The proposed algorithms have also been implemented on three benchmark datasets and execution times are reported.","PeriodicalId":147099,"journal":{"name":"2017 Tenth International Conference on Contemporary Computing (IC3)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128527679","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Biometric analysis for the recognition of spider species according to their webs 根据蜘蛛网识别蜘蛛种类的生物特征分析
Pub Date : 2017-08-01 DOI: 10.1109/IC3.2017.8284286
David Batista-Plaza, C. Travieso-González, M. Dutta, Anushikha Singh
This work presents a biometric approach for spider identification based on transform domain and Support Vector Machines as classifier. The dataset is composed by 185 images of spider web. The goal of this work is to use the structure of spider web for identifying the kind of spider. The experiments were done using two different of segmentation blocks and the analysis of the whole and center of the spider web. The best accuracy is reached after to run the different combinations.
本文提出了一种基于变换域和支持向量机作为分类器的蜘蛛生物识别方法。该数据集由185张蜘蛛网图像组成。这项工作的目的是利用蜘蛛网的结构来识别蜘蛛的种类。实验采用了两种不同的分割块,对蜘蛛网的整体和中心进行了分析。经过不同组合的运行,达到了最佳的精度。
{"title":"Biometric analysis for the recognition of spider species according to their webs","authors":"David Batista-Plaza, C. Travieso-González, M. Dutta, Anushikha Singh","doi":"10.1109/IC3.2017.8284286","DOIUrl":"https://doi.org/10.1109/IC3.2017.8284286","url":null,"abstract":"This work presents a biometric approach for spider identification based on transform domain and Support Vector Machines as classifier. The dataset is composed by 185 images of spider web. The goal of this work is to use the structure of spider web for identifying the kind of spider. The experiments were done using two different of segmentation blocks and the analysis of the whole and center of the spider web. The best accuracy is reached after to run the different combinations.","PeriodicalId":147099,"journal":{"name":"2017 Tenth International Conference on Contemporary Computing (IC3)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129285178","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Meta-heuristic solution for relay nodes placement in constrained environment 约束环境下中继节点布置的元启发式解决方案
Pub Date : 2017-08-01 DOI: 10.1109/IC3.2017.8284337
Manish Kumar, V. Ranga
Wireless sensor networks equipped with tiny and low powered nodes are susceptible to failures due to harsh surroundings. The operation of sensors becomes quite difficult when obstacles are present in the deployment area. Due to these obstacles, restoration of lost connectivity in WSN is a quite challenging task as well as computational intensive. Therefore, we proposed a meta-heuristic solution for restoration of lost connectivity. We use alpha shapes to detect boundary and shape of obstacles. Further, Grey Wolf Optimizer (GWO) is used to optimize the relay nodes placement. Our proposed solution, named as Meta-Heuristic Solution for Relay Node Placement in Constrained Environment (MH-RNPCE), implement convex hull approach to restrict the area for deployment of relay nodes. The simulation results show that the performance of MH-RNPCE.
配备微小和低功率节点的无线传感器网络容易因恶劣环境而发生故障。当部署区域存在障碍物时,传感器的操作变得相当困难。由于这些障碍,在WSN中恢复失去的连通性是一项相当具有挑战性的任务,并且计算量很大。因此,我们提出了一种元启发式解决方案来恢复丢失的连接。我们使用alpha形状来检测障碍物的边界和形状。进一步,利用灰狼优化器(GWO)优化中继节点的布局。我们提出的解决方案被命名为约束环境中中继节点放置的元启发式解决方案(MH-RNPCE),该方案采用凸包方法来限制中继节点的部署面积。仿真结果表明了MH-RNPCE的性能。
{"title":"Meta-heuristic solution for relay nodes placement in constrained environment","authors":"Manish Kumar, V. Ranga","doi":"10.1109/IC3.2017.8284337","DOIUrl":"https://doi.org/10.1109/IC3.2017.8284337","url":null,"abstract":"Wireless sensor networks equipped with tiny and low powered nodes are susceptible to failures due to harsh surroundings. The operation of sensors becomes quite difficult when obstacles are present in the deployment area. Due to these obstacles, restoration of lost connectivity in WSN is a quite challenging task as well as computational intensive. Therefore, we proposed a meta-heuristic solution for restoration of lost connectivity. We use alpha shapes to detect boundary and shape of obstacles. Further, Grey Wolf Optimizer (GWO) is used to optimize the relay nodes placement. Our proposed solution, named as Meta-Heuristic Solution for Relay Node Placement in Constrained Environment (MH-RNPCE), implement convex hull approach to restrict the area for deployment of relay nodes. The simulation results show that the performance of MH-RNPCE.","PeriodicalId":147099,"journal":{"name":"2017 Tenth International Conference on Contemporary Computing (IC3)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124202983","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Sentiment analysis using relative prosody features 基于相对韵律特征的情感分析
Pub Date : 2017-08-01 DOI: 10.1109/IC3.2017.8284296
Harika Abburi, K. R. Alluri, A. Vuppala, Manish Shrivastava, S. Gangashetty
Recent improvement in usage of digital media has led people to share their opinions about specific entity through audio. In this paper, an approach to detect the sentiment of an online spoken reviews based on relative prosody features is presented. Most of the existing systems for audio based sentiment analysis use conventional audio features, but they are not problem specific features to extract the sentiment. In this work, relative prosody features are extracted from normal and stressed regions of audio signal to detect the sentiment. Stressed regions are identified using the strength of excitation. Support Vector Machine (SVM) and Gaussian Mixture Model (GMM) classifiers are used to build the sentiment models. MOUD database is used for the proposed study. Experimental results show that, the rate of detecting the sentiment is improved with relative prosody features compared with the prosody and Mel Frequency Cepstral Coefficients (MFCC) because the relative prosody features has more sentiment specific discrimination compared to prosody features.
最近数字媒体使用的改善使得人们通过音频来分享他们对特定实体的看法。本文提出了一种基于相对韵律特征的在线语音评论情感检测方法。现有的基于音频的情感分析系统大多使用传统的音频特征,但它们不是提取情感的问题特定特征。在这项工作中,从音频信号的正常和重读区域提取相对韵律特征来检测情感。用激励强度来确定受力区域。使用支持向量机(SVM)和高斯混合模型(GMM)分类器构建情感模型。本研究采用mod数据库。实验结果表明,相对韵律特征比韵律特征和Mel频率倒谱系数(MFCC)更能提高情感检测率,因为相对韵律特征比韵律特征更能识别情感。
{"title":"Sentiment analysis using relative prosody features","authors":"Harika Abburi, K. R. Alluri, A. Vuppala, Manish Shrivastava, S. Gangashetty","doi":"10.1109/IC3.2017.8284296","DOIUrl":"https://doi.org/10.1109/IC3.2017.8284296","url":null,"abstract":"Recent improvement in usage of digital media has led people to share their opinions about specific entity through audio. In this paper, an approach to detect the sentiment of an online spoken reviews based on relative prosody features is presented. Most of the existing systems for audio based sentiment analysis use conventional audio features, but they are not problem specific features to extract the sentiment. In this work, relative prosody features are extracted from normal and stressed regions of audio signal to detect the sentiment. Stressed regions are identified using the strength of excitation. Support Vector Machine (SVM) and Gaussian Mixture Model (GMM) classifiers are used to build the sentiment models. MOUD database is used for the proposed study. Experimental results show that, the rate of detecting the sentiment is improved with relative prosody features compared with the prosody and Mel Frequency Cepstral Coefficients (MFCC) because the relative prosody features has more sentiment specific discrimination compared to prosody features.","PeriodicalId":147099,"journal":{"name":"2017 Tenth International Conference on Contemporary Computing (IC3)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121929936","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sentiment analysis of text using deep convolution neural networks 基于深度卷积神经网络的文本情感分析
Pub Date : 2017-08-01 DOI: 10.1109/IC3.2017.8284327
Anmol Chachra, Pulkit Mehndiratta, Mohit Gupta
Sentiment analysis has been one of the most researched topics in Machine learning. The roots of sentiment analysis are in studies on public opinion analysis at the start of 20th century, but the outbreak of computer-based sentiment analysis only occurred with the availability of subjective text in Web. The task of generating effective sentence model that captures both syntactic and semantic relations has been the primary goal to make better sentiment analyzers. In this paper, we harness the power of deep convolution neural networks (DCNN) to model sentences and perform sentiment analysis. This approach automates the whole process otherwise done using advance NLP techniques. It is a modular approach analyzing syntactic and context based relation from word level to phrase level to sentence level and then to document level. Such approach not only stands outs in terms of better classification, it also fits the concept of transfer learning. We have achieved an accuracy of 80.69% using this technique and further working on the enhancement and refinement of this approach.
情感分析一直是机器学习中研究最多的课题之一。情感分析的根源是20世纪初对民意分析的研究,但基于计算机的情感分析的爆发是随着网络上的主观文本的出现。生成能够捕获句法和语义关系的有效句子模型一直是构建更好的情感分析工具的主要目标。在本文中,我们利用深度卷积神经网络(DCNN)的力量来建模句子并进行情感分析。这种方法使整个过程自动化,否则使用先进的NLP技术完成。它是一种从词级到短语级到句子级再到文档级分析句法和上下文关系的模块化方法。这种方法不仅在更好的分类方面脱颖而出,而且也符合迁移学习的概念。我们已经使用该技术实现了80.69%的准确率,并进一步对该方法进行了改进和改进。
{"title":"Sentiment analysis of text using deep convolution neural networks","authors":"Anmol Chachra, Pulkit Mehndiratta, Mohit Gupta","doi":"10.1109/IC3.2017.8284327","DOIUrl":"https://doi.org/10.1109/IC3.2017.8284327","url":null,"abstract":"Sentiment analysis has been one of the most researched topics in Machine learning. The roots of sentiment analysis are in studies on public opinion analysis at the start of 20th century, but the outbreak of computer-based sentiment analysis only occurred with the availability of subjective text in Web. The task of generating effective sentence model that captures both syntactic and semantic relations has been the primary goal to make better sentiment analyzers. In this paper, we harness the power of deep convolution neural networks (DCNN) to model sentences and perform sentiment analysis. This approach automates the whole process otherwise done using advance NLP techniques. It is a modular approach analyzing syntactic and context based relation from word level to phrase level to sentence level and then to document level. Such approach not only stands outs in terms of better classification, it also fits the concept of transfer learning. We have achieved an accuracy of 80.69% using this technique and further working on the enhancement and refinement of this approach.","PeriodicalId":147099,"journal":{"name":"2017 Tenth International Conference on Contemporary Computing (IC3)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134205367","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
Region and feature matching based vehicle tracking for accident detection 基于区域和特征匹配的车辆跟踪事故检测
Pub Date : 2017-08-01 DOI: 10.1109/IC3.2017.8284322
Abhinav Saini, S. Suregaonkar, Neena Gupta, V. Karar, Shashi Poddar
Intelligent traffic monitoring using video surveillance is one of the most important aspects in administering a modern smart city. A recent growth towards machine learning and computer vision techniques has provided an added impetus towards this growth. In this paper, an image processing based vehicle tracking technique is developed that does not require background subtraction process to be applied for extracting the region of interest. Instead, a hybrid of feature detection and region matching approach is suggested in this article, which helps in estimating vehicle trajectory over consequent frames. Later, the tracked path is monitored for the occurrence of any specific event while the vehicle passes through an intersection. The proposed scheme is found to work promisingly on the real world dataset and is able to detect the occurrence of an accident between two vehicles.
基于视频监控的智能交通监控是管理现代智慧城市的重要方面之一。最近机器学习和计算机视觉技术的发展为这一增长提供了额外的动力。本文提出了一种基于图像处理的车辆跟踪技术,该技术不需要使用背景减除处理来提取感兴趣区域。相反,本文提出了一种混合特征检测和区域匹配方法,这有助于估计后续帧上的车辆轨迹。随后,当车辆通过十字路口时,跟踪的路径将被监控是否发生任何特定事件。该方案在实际数据集上工作良好,能够检测两辆车之间发生的事故。
{"title":"Region and feature matching based vehicle tracking for accident detection","authors":"Abhinav Saini, S. Suregaonkar, Neena Gupta, V. Karar, Shashi Poddar","doi":"10.1109/IC3.2017.8284322","DOIUrl":"https://doi.org/10.1109/IC3.2017.8284322","url":null,"abstract":"Intelligent traffic monitoring using video surveillance is one of the most important aspects in administering a modern smart city. A recent growth towards machine learning and computer vision techniques has provided an added impetus towards this growth. In this paper, an image processing based vehicle tracking technique is developed that does not require background subtraction process to be applied for extracting the region of interest. Instead, a hybrid of feature detection and region matching approach is suggested in this article, which helps in estimating vehicle trajectory over consequent frames. Later, the tracked path is monitored for the occurrence of any specific event while the vehicle passes through an intersection. The proposed scheme is found to work promisingly on the real world dataset and is able to detect the occurrence of an accident between two vehicles.","PeriodicalId":147099,"journal":{"name":"2017 Tenth International Conference on Contemporary Computing (IC3)","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132908015","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Authorship attribution for textual data on online social networks 在线社交网络上文本数据的作者归属
Pub Date : 2017-08-01 DOI: 10.1109/IC3.2017.8284311
Ritu Banga, Pulkit Mehndiratta
Authorship Attribution, (AA) is a process of determining a particular document is written by which author among a list of suspected authors. Authorship attribution has been the problem from last six decades; when there were handwritten documents needed to be identified for the genuine author. Due to the technology advancement and increase in cybercrime and unlawful activities, this problem of AA becomes forth most important to trace out the author behind online messages. Over the past, many years research has been conducted to attribute the authorship of an author on the basis of their writing style as all authors possess different distinctiveness while writing a piece of document. This paper presents a comparative study of various machine learning approaches on different feature sets for authorship attribution on short text. The Twitter dataset has been used for comparison with varying sample size of a dataset of 10 prolific authors with various combinations of feature sets. The significance and impact of combinations of features while inferring different stylometric features has been reflected. The results of different approaches are compared based on their accuracy and precision values.
作者归属(AA)是一个确定特定文档是由可疑作者列表中的哪个作者撰写的过程。作者归属问题在过去的六十年中一直存在;当有手写的文件需要鉴定为真正的作者。由于技术的进步和网络犯罪和非法活动的增加,追查网络信息背后的作者成为AA问题的第四个最重要的问题。在过去的许多年里,人们进行了研究,根据作者的写作风格来确定作者的身份,因为所有作者在撰写一份文件时都具有不同的独特性。本文针对短文本作者归属的不同特征集,对不同的机器学习方法进行了比较研究。Twitter数据集被用于与10个具有不同特征集组合的多产作者的数据集的不同样本量进行比较。在推断不同文体特征时,特征组合的意义和影响得到了体现。比较了不同方法的精度和精密度值。
{"title":"Authorship attribution for textual data on online social networks","authors":"Ritu Banga, Pulkit Mehndiratta","doi":"10.1109/IC3.2017.8284311","DOIUrl":"https://doi.org/10.1109/IC3.2017.8284311","url":null,"abstract":"Authorship Attribution, (AA) is a process of determining a particular document is written by which author among a list of suspected authors. Authorship attribution has been the problem from last six decades; when there were handwritten documents needed to be identified for the genuine author. Due to the technology advancement and increase in cybercrime and unlawful activities, this problem of AA becomes forth most important to trace out the author behind online messages. Over the past, many years research has been conducted to attribute the authorship of an author on the basis of their writing style as all authors possess different distinctiveness while writing a piece of document. This paper presents a comparative study of various machine learning approaches on different feature sets for authorship attribution on short text. The Twitter dataset has been used for comparison with varying sample size of a dataset of 10 prolific authors with various combinations of feature sets. The significance and impact of combinations of features while inferring different stylometric features has been reflected. The results of different approaches are compared based on their accuracy and precision values.","PeriodicalId":147099,"journal":{"name":"2017 Tenth International Conference on Contemporary Computing (IC3)","volume":"71 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133618849","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
CPLAG: Efficient plagiarism detection using bitwise operations CPLAG:采用位操作的高效剽窃检测
Pub Date : 2017-08-01 DOI: 10.1109/IC3.2017.8284313
Shikha Jain, Parmeet Kaur, M. Goyal, G. Dhanalekshmi
Source code plagiarism in an academic environment is a serious concern of faculties. The paper presents an efficient plagiarism detection tool, CPLAG, for C programming language codes. The tool assesses the structure of the C programs based on a set of attributes and performs a binary encoding of the C code statements. Subsequently, it utilizes computationally inexpensive bitwise operations to detect similarity between the given C programs. The design of CPLAG considers the commonly used techniques to avoid detection of plagiarism for delivering an efficient performance. Moreover, it avoids the extensive computations as used by existing tools for plagiarism detection. Experiment results indicate that CPLAG can detect plagiarism with accuracy. The tool finds application in academic institutions for fair and efficient automatic evaluation and grading of programming assignments.
源代码抄袭在学术环境是一个严重关注的院系。本文提出了一种高效的C语言代码抄袭检测工具CPLAG。该工具根据一组属性评估C程序的结构,并对C代码语句执行二进制编码。随后,它利用计算成本低廉的按位操作来检测给定C程序之间的相似性。CPLAG的设计考虑了避免抄袭检测的常用技术,提供了高效的性能。此外,它避免了现有抄袭检测工具所使用的大量计算。实验结果表明CPLAG能够准确地检测出抄袭。该工具在学术机构中用于公平有效的编程作业自动评估和评分。
{"title":"CPLAG: Efficient plagiarism detection using bitwise operations","authors":"Shikha Jain, Parmeet Kaur, M. Goyal, G. Dhanalekshmi","doi":"10.1109/IC3.2017.8284313","DOIUrl":"https://doi.org/10.1109/IC3.2017.8284313","url":null,"abstract":"Source code plagiarism in an academic environment is a serious concern of faculties. The paper presents an efficient plagiarism detection tool, CPLAG, for C programming language codes. The tool assesses the structure of the C programs based on a set of attributes and performs a binary encoding of the C code statements. Subsequently, it utilizes computationally inexpensive bitwise operations to detect similarity between the given C programs. The design of CPLAG considers the commonly used techniques to avoid detection of plagiarism for delivering an efficient performance. Moreover, it avoids the extensive computations as used by existing tools for plagiarism detection. Experiment results indicate that CPLAG can detect plagiarism with accuracy. The tool finds application in academic institutions for fair and efficient automatic evaluation and grading of programming assignments.","PeriodicalId":147099,"journal":{"name":"2017 Tenth International Conference on Contemporary Computing (IC3)","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127241034","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
2017 Tenth International Conference on Contemporary Computing (IC3)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1