Pub Date : 2014-12-01DOI: 10.1109/PADSW.2014.7097809
Waqaas Munawar, Heba Khdr, Santiago Pagani, M. Shafique, Jian-Jia Chen, J. Henkel
The number and diversity of cores in on-chip systems is increasing rapidly. However, due to the Thermal Design Power (TDP) constraint, it is not possible to continuously operate all cores at the same time. Exceeding the TDP constraint may activate the Dynamic Thermal Management (DTM) to ensure thermal stability. Such hardware based closed-loop safeguards pose a big challenge in using many-core chips for real-time tasks. Managing the worst-case peak power usage of a chip can help toward resolving this issue. We present a scheme to minimize the peak power usage for frame-based and periodic real-time tasks on many-core processors by scheduling the sleep cycles for each active core and introduce the concept of a sufficient test for peak power consumption for task feasibility. We consider both inter-task and inter-core diversity in terms of power usage and present computationally efficient algorithms for peak power minimization for these cases, i.e., a special case of “homogeneous tasks on homogeneous cores” to the general case of “heterogeneous tasks on heterogeneous cores”. We evaluate our solution through extensive simulations using the 48-core SCC platform and gem5 architecture simulator. Our simulation results show the efficacy of our scheme.
{"title":"Peak Power Management for scheduling real-time tasks on heterogeneous many-core systems","authors":"Waqaas Munawar, Heba Khdr, Santiago Pagani, M. Shafique, Jian-Jia Chen, J. Henkel","doi":"10.1109/PADSW.2014.7097809","DOIUrl":"https://doi.org/10.1109/PADSW.2014.7097809","url":null,"abstract":"The number and diversity of cores in on-chip systems is increasing rapidly. However, due to the Thermal Design Power (TDP) constraint, it is not possible to continuously operate all cores at the same time. Exceeding the TDP constraint may activate the Dynamic Thermal Management (DTM) to ensure thermal stability. Such hardware based closed-loop safeguards pose a big challenge in using many-core chips for real-time tasks. Managing the worst-case peak power usage of a chip can help toward resolving this issue. We present a scheme to minimize the peak power usage for frame-based and periodic real-time tasks on many-core processors by scheduling the sleep cycles for each active core and introduce the concept of a sufficient test for peak power consumption for task feasibility. We consider both inter-task and inter-core diversity in terms of power usage and present computationally efficient algorithms for peak power minimization for these cases, i.e., a special case of “homogeneous tasks on homogeneous cores” to the general case of “heterogeneous tasks on heterogeneous cores”. We evaluate our solution through extensive simulations using the 48-core SCC platform and gem5 architecture simulator. Our simulation results show the efficacy of our scheme.","PeriodicalId":421740,"journal":{"name":"2014 20th IEEE International Conference on Parallel and Distributed Systems (ICPADS)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122036221","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-12-01DOI: 10.1109/PADSW.2014.7097798
Xudong Shi, Feiqi Su, J. Peir
Maintaining hardware cache coherence on future CMPs becomes increasingly important and difficult as the number of cores keeps accelerating in mainstream multicore chips. The simple snooping-bus coherence scheme is not suitable due to its limited scalability. The sparse coherence directory approach may incur extra cache invalidations due to a topological mismatch between the coherence directory and the directories of all cache modules. In this paper, we propose an innovative CMP coherence directory that has three important properties. First, the directory has a simple set-associative design with small associativity. The number of directory entries matches the total number of cache blocks. Second, an augmented Directory Lookaside Table (DLT) allows blocks to be displaced from their primary sets in the coherence directory for alleviating hot-set conflicts. Third, to avoid expensive presence bits, each copy of a block along with the located core ID occupies a separate directory entry. Performance evaluations based on multithreaded and multi-programmed workloads demonstrate significant advantages of the proposed CMP directory over directories with traditional set-associative or skewed associative designs.
{"title":"Directory Lookaside Table: Enabling scalable, low-conflict, many-core cache coherence directory","authors":"Xudong Shi, Feiqi Su, J. Peir","doi":"10.1109/PADSW.2014.7097798","DOIUrl":"https://doi.org/10.1109/PADSW.2014.7097798","url":null,"abstract":"Maintaining hardware cache coherence on future CMPs becomes increasingly important and difficult as the number of cores keeps accelerating in mainstream multicore chips. The simple snooping-bus coherence scheme is not suitable due to its limited scalability. The sparse coherence directory approach may incur extra cache invalidations due to a topological mismatch between the coherence directory and the directories of all cache modules. In this paper, we propose an innovative CMP coherence directory that has three important properties. First, the directory has a simple set-associative design with small associativity. The number of directory entries matches the total number of cache blocks. Second, an augmented Directory Lookaside Table (DLT) allows blocks to be displaced from their primary sets in the coherence directory for alleviating hot-set conflicts. Third, to avoid expensive presence bits, each copy of a block along with the located core ID occupies a separate directory entry. Performance evaluations based on multithreaded and multi-programmed workloads demonstrate significant advantages of the proposed CMP directory over directories with traditional set-associative or skewed associative designs.","PeriodicalId":421740,"journal":{"name":"2014 20th IEEE International Conference on Parallel and Distributed Systems (ICPADS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130698862","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-12-01DOI: 10.1109/PADSW.2014.7097822
Yuxi Wang, Zimu Zhou, Kaishun Wu
Due to the rapid growth of the smartphone applications and the fast development of the Wireless Local Area Networks (WLANs), numerous indoor location-based techniques have been proposed during the past several decades. Floorplan, which defines the structure and functionality of a specific indoor environment, becomes a hot topic nowadays. Conventional floorplan techniques leverage smartphone sensors combined with WiFi signals to construct the floorplan of a building. However, existing approaches with sensors cannot detect the shape of a corner, and the sensors cost huge amount of energy during the whole floorplan constructing process. In this paper, we propose a sensor-free approach to detect the shape of a certain corner leveraging WiFi signals without using sensors on smartphones. Instead of utilizing traditional wireless communication indicator Received Signal Strength (RSS), we leverage a finer-grained indicator Channel State Information (CSI) to detect the shape of a certain corner. The evaluation of our approach shows that CSI is more robust in sensor-free corner shape detection, and we have achieved over 85% detection accuracy in simulation and over 70% detection accuracy in real indoor experiments.
{"title":"Sensor-free corner shape detection by wireless networks","authors":"Yuxi Wang, Zimu Zhou, Kaishun Wu","doi":"10.1109/PADSW.2014.7097822","DOIUrl":"https://doi.org/10.1109/PADSW.2014.7097822","url":null,"abstract":"Due to the rapid growth of the smartphone applications and the fast development of the Wireless Local Area Networks (WLANs), numerous indoor location-based techniques have been proposed during the past several decades. Floorplan, which defines the structure and functionality of a specific indoor environment, becomes a hot topic nowadays. Conventional floorplan techniques leverage smartphone sensors combined with WiFi signals to construct the floorplan of a building. However, existing approaches with sensors cannot detect the shape of a corner, and the sensors cost huge amount of energy during the whole floorplan constructing process. In this paper, we propose a sensor-free approach to detect the shape of a certain corner leveraging WiFi signals without using sensors on smartphones. Instead of utilizing traditional wireless communication indicator Received Signal Strength (RSS), we leverage a finer-grained indicator Channel State Information (CSI) to detect the shape of a certain corner. The evaluation of our approach shows that CSI is more robust in sensor-free corner shape detection, and we have achieved over 85% detection accuracy in simulation and over 70% detection accuracy in real indoor experiments.","PeriodicalId":421740,"journal":{"name":"2014 20th IEEE International Conference on Parallel and Distributed Systems (ICPADS)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114132339","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-12-01DOI: 10.1109/PADSW.2014.7097926
Chuan-Hsiang Han, Yu-Tuan Lin
Monte Carlo simulations have become widely used in computational finance. Standard error is the basic notion to measure the quality of a Monte Carlo estimator, and the square of standard error is defined as the variance divided by the total number of simulations. Variance reduction methods have been developed as efficient algorithms by means of probabilistic analysis. GPU acceleration plays a crucial role of increasing the total number of simulations. We show that the total effect of combining variance reduction methods as efficient software algorithms with GPU acceleration as a parallel-computing hardware device can yield a tremendous speed up for financial applications such as evaluation of option prices and estimation of joint default probabilities.
{"title":"Accelerated variance reduction methods on GPU","authors":"Chuan-Hsiang Han, Yu-Tuan Lin","doi":"10.1109/PADSW.2014.7097926","DOIUrl":"https://doi.org/10.1109/PADSW.2014.7097926","url":null,"abstract":"Monte Carlo simulations have become widely used in computational finance. Standard error is the basic notion to measure the quality of a Monte Carlo estimator, and the square of standard error is defined as the variance divided by the total number of simulations. Variance reduction methods have been developed as efficient algorithms by means of probabilistic analysis. GPU acceleration plays a crucial role of increasing the total number of simulations. We show that the total effect of combining variance reduction methods as efficient software algorithms with GPU acceleration as a parallel-computing hardware device can yield a tremendous speed up for financial applications such as evaluation of option prices and estimation of joint default probabilities.","PeriodicalId":421740,"journal":{"name":"2014 20th IEEE International Conference on Parallel and Distributed Systems (ICPADS)","volume":"161 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114142295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-12-01DOI: 10.1109/PADSW.2014.7097898
Chia-Ping Tsai, Hung-Chang Hsiao
Present NoSQL databases are passive entities, where users proactively access the databases. While NoSQL databases are scalable due to their horizontal scale-out designs, data items stored in potentially very large databases are difficult to retrieve in terms of access delay, programmability and usability. In this paper, we advocate supporting events in NoSQL. By introducing publishers and subscribers to NoSQL, our proposed NoSQL data store is capable of delivering data items that users are interested in. Additionally, our NoSQL database decouples publishers and subscribers such that application developers can emphasize on data manipulation without paying attention to communications for delivering data. We formally discuss the design requirements for streaming in NoSQL, and present a prototype implementation that addresses the design issues. We also outline our ongoing works in this paper.
{"title":"Streaming in NoSQL","authors":"Chia-Ping Tsai, Hung-Chang Hsiao","doi":"10.1109/PADSW.2014.7097898","DOIUrl":"https://doi.org/10.1109/PADSW.2014.7097898","url":null,"abstract":"Present NoSQL databases are passive entities, where users proactively access the databases. While NoSQL databases are scalable due to their horizontal scale-out designs, data items stored in potentially very large databases are difficult to retrieve in terms of access delay, programmability and usability. In this paper, we advocate supporting events in NoSQL. By introducing publishers and subscribers to NoSQL, our proposed NoSQL data store is capable of delivering data items that users are interested in. Additionally, our NoSQL database decouples publishers and subscribers such that application developers can emphasize on data manipulation without paying attention to communications for delivering data. We formally discuss the design requirements for streaming in NoSQL, and present a prototype implementation that addresses the design issues. We also outline our ongoing works in this paper.","PeriodicalId":421740,"journal":{"name":"2014 20th IEEE International Conference on Parallel and Distributed Systems (ICPADS)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116426201","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-12-01DOI: 10.1109/PADSW.2014.7097802
Yongtao Zhou, Yuhui Deng, Junjie Xie
Data deduplication has been widely used at data backup system due to the significantly reduced requirements of storage capacity and network bandwidth. However, the performance of data deduplication gradually decreases with the growth of deduplicated data. This is because the volume of fingerprints grows significantly with the increase of backup data, and a large portion of fingerprints have to be stored on disk drives. This incurs frequent disk accesses to locate fingerprints and blocks the process of data deduplication. Furthermore, the fingerprints belonging to the same file may be discretely stored on disk drives. This generates random and small disk accesses, and results in significant performance degradation when the fingerprints are referred. Additionally, a single fingerprint may appear only once during a backup process. This results in very low cache hit ratio due to lacking temporal locality. This paper proposes to employ file similarity to enhance the fingerprint prefetching, thus improving the cache hit ratio and the performance of data deduplication. Furthermore, the fingerprints are arranged sequently in terms of the backup data stream to maintain the locality and promote the performance. Experimental results demonstrate that the proposed idea can effectively reduce the number of fingerprint accesses going to disk drives, decrease the query overhead of fingerprints, thus significantly alleviating the disk bottleneck of data deduplication.
{"title":"Leverage similarity and locality to enhance fingerprint prefetching of data deduplication","authors":"Yongtao Zhou, Yuhui Deng, Junjie Xie","doi":"10.1109/PADSW.2014.7097802","DOIUrl":"https://doi.org/10.1109/PADSW.2014.7097802","url":null,"abstract":"Data deduplication has been widely used at data backup system due to the significantly reduced requirements of storage capacity and network bandwidth. However, the performance of data deduplication gradually decreases with the growth of deduplicated data. This is because the volume of fingerprints grows significantly with the increase of backup data, and a large portion of fingerprints have to be stored on disk drives. This incurs frequent disk accesses to locate fingerprints and blocks the process of data deduplication. Furthermore, the fingerprints belonging to the same file may be discretely stored on disk drives. This generates random and small disk accesses, and results in significant performance degradation when the fingerprints are referred. Additionally, a single fingerprint may appear only once during a backup process. This results in very low cache hit ratio due to lacking temporal locality. This paper proposes to employ file similarity to enhance the fingerprint prefetching, thus improving the cache hit ratio and the performance of data deduplication. Furthermore, the fingerprints are arranged sequently in terms of the backup data stream to maintain the locality and promote the performance. Experimental results demonstrate that the proposed idea can effectively reduce the number of fingerprint accesses going to disk drives, decrease the query overhead of fingerprints, thus significantly alleviating the disk bottleneck of data deduplication.","PeriodicalId":421740,"journal":{"name":"2014 20th IEEE International Conference on Parallel and Distributed Systems (ICPADS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114424175","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-12-01DOI: 10.1109/PADSW.2014.7097922
Xiaowei Shen, Fenglong Song, Haibo Meng, Shuqian An, Zhimin Zhang
Memory requests in many-core systems are interleaved with each other and the locality of many-core systems decreases heavily. Page policies in traditional single core systems are not effective when it comes to many-core systems, because the open-page policy needs much locality of memory requests and the close-page policy takes no advantage of the remaining locality of many-core systems. There are some related memory page management policies, but their high complexity makes them unsuitable to many-core systems. They either need too much modification in operating systems or have large area and power overhead. To overcome these shortcomings of current page policies, in this paper, we propose the row based page policy, that is, RBPP, for the many-core systems, which tracks the row addresses of memory requests to each bank and uses row addresses as the indicator to decide whether or not to close the row buffer when the active memory request finished. We evaluate the proposed RBPP via Gem5 and DRAMSim2, and the results show that row based page policy can decrease the average memory latency by 14.7% and 4.0% over the open-page policy and the close-page policy, respectively. And the area overhead of row based page policy is decreased by 91.4 % and 91.5% over access based page policy and two-level predictor page policy, respectively.
{"title":"RBPP: A row based DRAM page policy for the many-core era","authors":"Xiaowei Shen, Fenglong Song, Haibo Meng, Shuqian An, Zhimin Zhang","doi":"10.1109/PADSW.2014.7097922","DOIUrl":"https://doi.org/10.1109/PADSW.2014.7097922","url":null,"abstract":"Memory requests in many-core systems are interleaved with each other and the locality of many-core systems decreases heavily. Page policies in traditional single core systems are not effective when it comes to many-core systems, because the open-page policy needs much locality of memory requests and the close-page policy takes no advantage of the remaining locality of many-core systems. There are some related memory page management policies, but their high complexity makes them unsuitable to many-core systems. They either need too much modification in operating systems or have large area and power overhead. To overcome these shortcomings of current page policies, in this paper, we propose the row based page policy, that is, RBPP, for the many-core systems, which tracks the row addresses of memory requests to each bank and uses row addresses as the indicator to decide whether or not to close the row buffer when the active memory request finished. We evaluate the proposed RBPP via Gem5 and DRAMSim2, and the results show that row based page policy can decrease the average memory latency by 14.7% and 4.0% over the open-page policy and the close-page policy, respectively. And the area overhead of row based page policy is decreased by 91.4 % and 91.5% over access based page policy and two-level predictor page policy, respectively.","PeriodicalId":421740,"journal":{"name":"2014 20th IEEE International Conference on Parallel and Distributed Systems (ICPADS)","volume":"64 7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134221356","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-12-01DOI: 10.1109/PADSW.2014.7097897
J. Brito, Aleteia P. F. Araujo
This paper describes a model which aims to estimate the size of a cluster running Hadoop framework for the processing of large datasets at a given timeframe. As main contributions it denes (i) a light layer of optimization for MapReduce jobs, (ii) presents a model to estimate the size cluster for a Hadoop framework and (iii) performs tests using a real environment - the Amazon Elastic MapReduce. The proposed approach works with the MapReduce to dene the main configuration parameters and determines computational resources of hosts in the cluster in order to meet the desired runtime for the requirements of a given workload requirement. Thus, the results show that the proposed model is able to avoid to over-allocation or sub-allocation of computing resources on a Hadoop cluster.
{"title":"Model to estimate the size of a Hadoop cluster - HCEm","authors":"J. Brito, Aleteia P. F. Araujo","doi":"10.1109/PADSW.2014.7097897","DOIUrl":"https://doi.org/10.1109/PADSW.2014.7097897","url":null,"abstract":"This paper describes a model which aims to estimate the size of a cluster running Hadoop framework for the processing of large datasets at a given timeframe. As main contributions it denes (i) a light layer of optimization for MapReduce jobs, (ii) presents a model to estimate the size cluster for a Hadoop framework and (iii) performs tests using a real environment - the Amazon Elastic MapReduce. The proposed approach works with the MapReduce to dene the main configuration parameters and determines computational resources of hosts in the cluster in order to meet the desired runtime for the requirements of a given workload requirement. Thus, the results show that the proposed model is able to avoid to over-allocation or sub-allocation of computing resources on a Hadoop cluster.","PeriodicalId":421740,"journal":{"name":"2014 20th IEEE International Conference on Parallel and Distributed Systems (ICPADS)","volume":"99 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133627141","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-12-01DOI: 10.1109/PADSW.2014.7097886
Yuki Yagi, Naofumi Kitsunezaki, H. Saito, Y. Tobe
In this research, we designed and implemented Real-World File System (RWFS), which can manage files as if we can put them onto or pick them up from the places of the real world. RWFS regards the places of the real world as directories of the file system by associating a directory with a place. We create directories called Real-World Directory (RWD) which forms a hierarchical structure to reflect the natural property of places. In addition to the conventional access rights of read, write, and execute as implemented in other file systems, RWFS accommodates utilizing location information of the target user in access rights; RWFS can decide whether or not the user can access to a particular file or directory based on the user's location. Therefore, accessible files for a user change depending on the user's location. This mechanism enables creating information that can be read or written by users who physically stay at a particular place. We evaluated this system by measuring turnaround time to operate the file system together with simulation.
{"title":"RWFS: Design and implementation of file system executing access control based on user's location","authors":"Yuki Yagi, Naofumi Kitsunezaki, H. Saito, Y. Tobe","doi":"10.1109/PADSW.2014.7097886","DOIUrl":"https://doi.org/10.1109/PADSW.2014.7097886","url":null,"abstract":"In this research, we designed and implemented Real-World File System (RWFS), which can manage files as if we can put them onto or pick them up from the places of the real world. RWFS regards the places of the real world as directories of the file system by associating a directory with a place. We create directories called Real-World Directory (RWD) which forms a hierarchical structure to reflect the natural property of places. In addition to the conventional access rights of read, write, and execute as implemented in other file systems, RWFS accommodates utilizing location information of the target user in access rights; RWFS can decide whether or not the user can access to a particular file or directory based on the user's location. Therefore, accessible files for a user change depending on the user's location. This mechanism enables creating information that can be read or written by users who physically stay at a particular place. We evaluated this system by measuring turnaround time to operate the file system together with simulation.","PeriodicalId":421740,"journal":{"name":"2014 20th IEEE International Conference on Parallel and Distributed Systems (ICPADS)","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133995653","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-12-01DOI: 10.1109/PADSW.2014.7097892
Qiang Li, Jie Yu, Zhoujun Li
Kad is the most popular P2P file sharing system. Monitoring Kad peers' lookup traffic is an important work for the analysis and optimization of Peer-to-Peer (P2P) network. During the monitoring process, we find that the peer's status significantly influences the monitoring results. Each lookup action changes the searching peer's routing table status, and it may break the monitoring process. In this paper, we analyze the changes in the routing table to verify its effect on the monitoring process. If the distance between the target ID and searcher's Kad ID is in within a certain critical range, previous searches may cause future searches to fail with high probability. We estimate the boundary of this critical range. The experiments performed on eMule shows that such a critical range exist, and that deploying more than 1024 IP addresses cannot help to improve the success rate of the monitoring process.
{"title":"Routing table status influence of Monitoring Kad","authors":"Qiang Li, Jie Yu, Zhoujun Li","doi":"10.1109/PADSW.2014.7097892","DOIUrl":"https://doi.org/10.1109/PADSW.2014.7097892","url":null,"abstract":"Kad is the most popular P2P file sharing system. Monitoring Kad peers' lookup traffic is an important work for the analysis and optimization of Peer-to-Peer (P2P) network. During the monitoring process, we find that the peer's status significantly influences the monitoring results. Each lookup action changes the searching peer's routing table status, and it may break the monitoring process. In this paper, we analyze the changes in the routing table to verify its effect on the monitoring process. If the distance between the target ID and searcher's Kad ID is in within a certain critical range, previous searches may cause future searches to fail with high probability. We estimate the boundary of this critical range. The experiments performed on eMule shows that such a critical range exist, and that deploying more than 1024 IP addresses cannot help to improve the success rate of the monitoring process.","PeriodicalId":421740,"journal":{"name":"2014 20th IEEE International Conference on Parallel and Distributed Systems (ICPADS)","volume":"61 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116594600","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}