{"title":"Job starvation avoidance with alleviation of data skewness in Big Data infrastructure","authors":"Sankari Subbiah, S. Mala, S. Nayagam","doi":"10.1109/ICCCT2.2017.7972264","DOIUrl":null,"url":null,"abstract":"During the age of rush in the need for big data, Hadoop is a postulate or cloud-based platform that has been heavily encouraged for all solutions in the business world's big data problems. Parallel execution of jobs consists of large data sets is done through map reduce in the hadoop cluster. The completion of job time will depend on the slowest running task in the job. The entire job is extended if one particular job takes longer time to finish and it is done by the delayer. An inequality in the measure of data allocated to each individual task is referred to as Data skewness. An efficient dynamic data splitting approach on Hadoop called the Hybrid scheduler who monitors the samples while running batch jobs and allocates resources to slaves depending on the complexity of data and the time taken for processing. In this paper, the effectiveness of web swarming is showcased using hadoop eliminating Distributed Denial of Service (DDoS) attack detection scenarios in the Web servers. Query processing is done through Map Reduce in traditional Hadoop clusters and is replaced by the proposed Block chain query processing algorithm. Thereby improvise the execution time of the assigned task in the proposed system to mitigate the data skewness. The main aim of this paper is to avoid job starvation thus minimizing the response time efficiently during the process and mitigating data skewness in existing system.","PeriodicalId":445567,"journal":{"name":"2017 2nd International Conference on Computing and Communications Technologies (ICCCT)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2017-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 2nd International Conference on Computing and Communications Technologies (ICCCT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCCT2.2017.7972264","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3
Abstract
During the age of rush in the need for big data, Hadoop is a postulate or cloud-based platform that has been heavily encouraged for all solutions in the business world's big data problems. Parallel execution of jobs consists of large data sets is done through map reduce in the hadoop cluster. The completion of job time will depend on the slowest running task in the job. The entire job is extended if one particular job takes longer time to finish and it is done by the delayer. An inequality in the measure of data allocated to each individual task is referred to as Data skewness. An efficient dynamic data splitting approach on Hadoop called the Hybrid scheduler who monitors the samples while running batch jobs and allocates resources to slaves depending on the complexity of data and the time taken for processing. In this paper, the effectiveness of web swarming is showcased using hadoop eliminating Distributed Denial of Service (DDoS) attack detection scenarios in the Web servers. Query processing is done through Map Reduce in traditional Hadoop clusters and is replaced by the proposed Block chain query processing algorithm. Thereby improvise the execution time of the assigned task in the proposed system to mitigate the data skewness. The main aim of this paper is to avoid job starvation thus minimizing the response time efficiently during the process and mitigating data skewness in existing system.