{"title":"内存文本数据分析的多gpu框架","authors":"P. K. Chong, E. Karuppiah, K. Yong","doi":"10.1109/WAINA.2013.238","DOIUrl":null,"url":null,"abstract":"Current application of GPU processors for parallel computing tasks show excellent results in terms of speed-ups compared to CPU processors. However, there is no existing framework that enables automatic distribution of data and processing across multiple GPUs, modularity of kernel design, and efficient co-usage of CPU and GPU processors. All these elements are necessary conditions to enable users to easily perform 'Big Data' analysis, and to create their own modules for their desired processing functionality. We propose a framework for in-memory 'Big Text Data' analytics that provides mechanisms for automatic data segmentation, distribution, execution, and result retrieval across multiple cards (CPU, GPU & FPGA) and machines, and a modular design for easy addition of new GPU kernels. The architecture and components of the framework such as multi-card data distribution and execution, data structures for efficient memory access, algorithms for parallel GPU computation, and result retrieval are described in detail, and some of the kernels in the framework are evaluated using Big Data versus multi-core CPUs to demonstrate the performance and feasibility of using it for 'Big Data' analytics, providing alternative and cheaper HPC solution.","PeriodicalId":359251,"journal":{"name":"2013 27th International Conference on Advanced Information Networking and Applications Workshops","volume":"13 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2013-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":"{\"title\":\"A Multi-GPU Framework for In-Memory Text Data Analytics\",\"authors\":\"P. K. Chong, E. Karuppiah, K. Yong\",\"doi\":\"10.1109/WAINA.2013.238\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Current application of GPU processors for parallel computing tasks show excellent results in terms of speed-ups compared to CPU processors. However, there is no existing framework that enables automatic distribution of data and processing across multiple GPUs, modularity of kernel design, and efficient co-usage of CPU and GPU processors. All these elements are necessary conditions to enable users to easily perform 'Big Data' analysis, and to create their own modules for their desired processing functionality. We propose a framework for in-memory 'Big Text Data' analytics that provides mechanisms for automatic data segmentation, distribution, execution, and result retrieval across multiple cards (CPU, GPU & FPGA) and machines, and a modular design for easy addition of new GPU kernels. The architecture and components of the framework such as multi-card data distribution and execution, data structures for efficient memory access, algorithms for parallel GPU computation, and result retrieval are described in detail, and some of the kernels in the framework are evaluated using Big Data versus multi-core CPUs to demonstrate the performance and feasibility of using it for 'Big Data' analytics, providing alternative and cheaper HPC solution.\",\"PeriodicalId\":359251,\"journal\":{\"name\":\"2013 27th International Conference on Advanced Information Networking and Applications Workshops\",\"volume\":\"13 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2013-03-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"7\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2013 27th International Conference on Advanced Information Networking and Applications Workshops\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/WAINA.2013.238\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2013 27th International Conference on Advanced Information Networking and Applications Workshops","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/WAINA.2013.238","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A Multi-GPU Framework for In-Memory Text Data Analytics
Current application of GPU processors for parallel computing tasks show excellent results in terms of speed-ups compared to CPU processors. However, there is no existing framework that enables automatic distribution of data and processing across multiple GPUs, modularity of kernel design, and efficient co-usage of CPU and GPU processors. All these elements are necessary conditions to enable users to easily perform 'Big Data' analysis, and to create their own modules for their desired processing functionality. We propose a framework for in-memory 'Big Text Data' analytics that provides mechanisms for automatic data segmentation, distribution, execution, and result retrieval across multiple cards (CPU, GPU & FPGA) and machines, and a modular design for easy addition of new GPU kernels. The architecture and components of the framework such as multi-card data distribution and execution, data structures for efficient memory access, algorithms for parallel GPU computation, and result retrieval are described in detail, and some of the kernels in the framework are evaluated using Big Data versus multi-core CPUs to demonstrate the performance and feasibility of using it for 'Big Data' analytics, providing alternative and cheaper HPC solution.