{"title":"Map/Reduce范式的Petri网形式化以优化性能-成本权衡","authors":"M. Ruiz, Javier L. Calleja, D. Cazorla","doi":"10.1109/Trustcom.2015.617","DOIUrl":null,"url":null,"abstract":"Nowadays, the world around us is built up on more and more unstructured data every day. However, performing a longitudinal analysis of these data becomes a Big-Data problem that cannot be tackled with traditional tools, storage or processing infrastructures. One of the main contributions to address this matter has been the Hadoop framework (which implements the Map/Reduce paradigm), especially when used in conjunction with Cloud computing environments. This paper presents a formalization of the Map/Reduce paradigm which is used to evaluate performance parameters and make a trade-off analysis of the number of workers versus processing time and resource cost. We have used Prioritised -- Timed Coloured Petri Nets to obtain complete and unambiguous models of the system behaviour as well as CPNTools to evaluate the correctness of the system using state space exploration and for performance evaluation. The resulting formal model is evaluated with a real social media data Hadoop-based application and it is validated by carrying out experiments on a real private Cloud environment. Results show that the proposed model enables to determine in advance both the performance of a Map/Reduce-based application within Cloud environments and the best performance-cost agreement.","PeriodicalId":277092,"journal":{"name":"2015 IEEE Trustcom/BigDataSE/ISPA","volume":"124 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2015-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"11","resultStr":"{\"title\":\"Petri Nets Formalization of Map/Reduce Paradigm to Optimise the Performance-Cost Tradeoff\",\"authors\":\"M. Ruiz, Javier L. Calleja, D. Cazorla\",\"doi\":\"10.1109/Trustcom.2015.617\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Nowadays, the world around us is built up on more and more unstructured data every day. However, performing a longitudinal analysis of these data becomes a Big-Data problem that cannot be tackled with traditional tools, storage or processing infrastructures. One of the main contributions to address this matter has been the Hadoop framework (which implements the Map/Reduce paradigm), especially when used in conjunction with Cloud computing environments. This paper presents a formalization of the Map/Reduce paradigm which is used to evaluate performance parameters and make a trade-off analysis of the number of workers versus processing time and resource cost. We have used Prioritised -- Timed Coloured Petri Nets to obtain complete and unambiguous models of the system behaviour as well as CPNTools to evaluate the correctness of the system using state space exploration and for performance evaluation. The resulting formal model is evaluated with a real social media data Hadoop-based application and it is validated by carrying out experiments on a real private Cloud environment. Results show that the proposed model enables to determine in advance both the performance of a Map/Reduce-based application within Cloud environments and the best performance-cost agreement.\",\"PeriodicalId\":277092,\"journal\":{\"name\":\"2015 IEEE Trustcom/BigDataSE/ISPA\",\"volume\":\"124 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2015-08-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"11\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2015 IEEE Trustcom/BigDataSE/ISPA\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/Trustcom.2015.617\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2015 IEEE Trustcom/BigDataSE/ISPA","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/Trustcom.2015.617","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Petri Nets Formalization of Map/Reduce Paradigm to Optimise the Performance-Cost Tradeoff
Nowadays, the world around us is built up on more and more unstructured data every day. However, performing a longitudinal analysis of these data becomes a Big-Data problem that cannot be tackled with traditional tools, storage or processing infrastructures. One of the main contributions to address this matter has been the Hadoop framework (which implements the Map/Reduce paradigm), especially when used in conjunction with Cloud computing environments. This paper presents a formalization of the Map/Reduce paradigm which is used to evaluate performance parameters and make a trade-off analysis of the number of workers versus processing time and resource cost. We have used Prioritised -- Timed Coloured Petri Nets to obtain complete and unambiguous models of the system behaviour as well as CPNTools to evaluate the correctness of the system using state space exploration and for performance evaluation. The resulting formal model is evaluated with a real social media data Hadoop-based application and it is validated by carrying out experiments on a real private Cloud environment. Results show that the proposed model enables to determine in advance both the performance of a Map/Reduce-based application within Cloud environments and the best performance-cost agreement.