首页 > 最新文献

2010 IEEE 16th International Conference on Parallel and Distributed Systems最新文献

英文 中文
Unifying Buffer Replacement and Prefetching with Data Migration for Heterogeneous Storage Devices 异构存储设备缓冲区替换和预取与数据迁移的统一
Pub Date : 2010-12-08 DOI: 10.1109/ICPADS.2010.103
Sehwan Lee, K. Koh, H. Bahn
With the good properties of NAND flash memory such as small size, shock resistance, and low-power consumption, large capacity SSD (Solid State Disk) is anticipated to replace hard disk in high-end systems. However, the cost of NAND flash memory is still high to substitute for hard disk entirely. Using hard disk and NAND flash memory together as secondary storage is an alternative solution to provide relatively low response time, large capacity, and reasonable cost. In this paper, we present a new buffer cache management scheme with data migration that is optimized to use both NAND flash memory and hard disk together as secondary storage. The proposed scheme has three salient features. First, it detects I/O access patterns from each storage, and allocates the buffer cache space for each pattern by computing marginal gain adaptively considering the I/O cost of storage. Second, it prefetches data selectively according to their access pattern and storage devices. Third, it moves the evicted data from the buffer cache to hard disk or NAND flash memory considering the access patterns of block references on the reclamation. Trace-driven simulations show that the proposed scheme improves the I/O performance significantly. It enhances the buffer cache hit ratio by up to 29.9% and reduces the total I/O elapsed time by up to 49.5% compared to the well-acknowledged UBM scheme.
由于NAND闪存具有体积小、耐冲击、低功耗等优点,大容量SSD (Solid State Disk)有望在高端系统中取代硬盘。然而,NAND闪存的成本仍然很高,无法完全取代硬盘。使用硬盘和NAND闪存作为辅助存储是一种替代解决方案,可以提供相对较低的响应时间、较大的容量和合理的成本。在本文中,我们提出了一种新的缓冲区高速缓存管理方案,该方案优化了NAND闪存和硬盘作为辅助存储的使用。提出的方案有三个显著特点。首先,它从每个存储中检测I/O访问模式,并根据存储的I/O成本自适应计算边际增益,为每个模式分配缓冲缓存空间。其次,根据数据的访问方式和存储设备选择性地预取数据。第三,考虑回收时块引用的访问模式,将被驱逐的数据从缓冲缓存移动到硬盘或NAND闪存。跟踪驱动仿真表明,该方案显著提高了I/O性能。与公认的UBM方案相比,它将缓冲区缓存命中率提高了29.9%,并将总I/O消耗时间减少了49.5%。
{"title":"Unifying Buffer Replacement and Prefetching with Data Migration for Heterogeneous Storage Devices","authors":"Sehwan Lee, K. Koh, H. Bahn","doi":"10.1109/ICPADS.2010.103","DOIUrl":"https://doi.org/10.1109/ICPADS.2010.103","url":null,"abstract":"With the good properties of NAND flash memory such as small size, shock resistance, and low-power consumption, large capacity SSD (Solid State Disk) is anticipated to replace hard disk in high-end systems. However, the cost of NAND flash memory is still high to substitute for hard disk entirely. Using hard disk and NAND flash memory together as secondary storage is an alternative solution to provide relatively low response time, large capacity, and reasonable cost. In this paper, we present a new buffer cache management scheme with data migration that is optimized to use both NAND flash memory and hard disk together as secondary storage. The proposed scheme has three salient features. First, it detects I/O access patterns from each storage, and allocates the buffer cache space for each pattern by computing marginal gain adaptively considering the I/O cost of storage. Second, it prefetches data selectively according to their access pattern and storage devices. Third, it moves the evicted data from the buffer cache to hard disk or NAND flash memory considering the access patterns of block references on the reclamation. Trace-driven simulations show that the proposed scheme improves the I/O performance significantly. It enhances the buffer cache hit ratio by up to 29.9% and reduces the total I/O elapsed time by up to 49.5% compared to the well-acknowledged UBM scheme.","PeriodicalId":365914,"journal":{"name":"2010 IEEE 16th International Conference on Parallel and Distributed Systems","volume":"226 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131445985","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
PGOS: An Architecture of a Personal Net Computing Platform PGOS:一种个人网络计算平台体系结构
Pub Date : 2010-12-08 DOI: 10.1109/ICPADS.2010.50
Jie Liu, Yongqiang Zou
The development of web has brought rich applications and services, giving users convenience, but also causing that the user’s information is locked and isolated, the users’ resource are disperse, and the operation granularity is not uniform, which ultimately harm the end-users. This paper presents PGOS,Personal Grid Operation System, a general-purpose software for controlled sharing of cross-domain resources in personal net computing [1]. It accesses disperse resources uniformly from web client in order to connect the information islands formed by the companies, and it provides uniform fine-grained sharing mechanism, in addition, we can build new applications by combining the integrated resources in PGOS. The article proposes PGOS Core to complete decentralized user authentication, authorization and access control, Funnel is used to abstract the resource and make decentralized resource discovery, simultaneously PGSML, the Personal Grid Service Markup Language, is put forward to construct PGOS applications.
web的发展给用户带来了丰富的应用和服务,给用户带来了便利,但也造成了用户信息的锁定和隔离、用户资源的分散、操作粒度的不统一,最终伤害了终端用户。本文介绍了PGOS (Personal Grid Operation System,个人网格操作系统),一个用于个人网络计算[1]中跨域资源可控共享的通用软件。它从web客户端统一访问分散的资源,将企业形成的信息孤岛连接起来,并提供统一的细粒度共享机制,此外,我们还可以通过整合PGOS中的资源来构建新的应用程序。本文提出了PGOS核心来完成去中心化的用户认证、授权和访问控制,利用Funnel对资源进行抽象和去中心化的资源发现,同时提出了PGSML——个人网格服务标记语言来构建PGOS应用。
{"title":"PGOS: An Architecture of a Personal Net Computing Platform","authors":"Jie Liu, Yongqiang Zou","doi":"10.1109/ICPADS.2010.50","DOIUrl":"https://doi.org/10.1109/ICPADS.2010.50","url":null,"abstract":"The development of web has brought rich applications and services, giving users convenience, but also causing that the user’s information is locked and isolated, the users’ resource are disperse, and the operation granularity is not uniform, which ultimately harm the end-users. This paper presents PGOS,Personal Grid Operation System, a general-purpose software for controlled sharing of cross-domain resources in personal net computing [1]. It accesses disperse resources uniformly from web client in order to connect the information islands formed by the companies, and it provides uniform fine-grained sharing mechanism, in addition, we can build new applications by combining the integrated resources in PGOS. The article proposes PGOS Core to complete decentralized user authentication, authorization and access control, Funnel is used to abstract the resource and make decentralized resource discovery, simultaneously PGSML, the Personal Grid Service Markup Language, is put forward to construct PGOS applications.","PeriodicalId":365914,"journal":{"name":"2010 IEEE 16th International Conference on Parallel and Distributed Systems","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115750166","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Utilizing RF Interference to Enable Private Estimation in RFID Systems 利用射频干扰实现RFID系统的私有估计
Pub Date : 2010-12-08 DOI: 10.1109/ICPADS.2010.106
Lei Yang, Jinsong Han, Yong Qi, Cheng Wang, Zhuo Li, Qingsong Yao, Ying Chen, Xiao Zhong
Counting or estimating the number of tags is crucial for RFID system. Researchers have proposed several fast cardinality estimation schemes to estimate the quantity of a batch of tags within a short time frame. Existing estimation schemes scarcely consider the privacy issue. Without effective protection, the adversary can utilize the responding signals to estimate the number of tags as accurate as the valid reader. To address this issue, we propose a novel privacy-preserving estimation scheme, termed as MEAS, which provides an active RF countermeasure against the estimation from invalid readers. MEAS comprises of two components, an Estimation Interference Device (EID) and two well-designed Interference Blanking Estimators (IBE). EID is deployed with the tags to actively generate interfering signals, which introduce sufficiently large estimation errors to invalid or malicious readers. Using a secret interference factor shared with EID, a valid reader can perform accurate estimation via two IBEs. Our theoretical analysis and simulation results show the effectiveness of MEAS. Meanwhile, MEAS can also maintain a high estimation accuracy using IBEs.
标签数量的计算或估计对RFID系统至关重要。为了在短时间内估计一批标签的数量,研究人员提出了几种快速基数估计方案。现有的估计方案很少考虑隐私问题。如果没有有效的保护,攻击者可以利用响应信号来估计标签的数量,与有效的读取器一样准确。为了解决这个问题,我们提出了一种新的隐私保护估计方案,称为MEAS,它提供了一种主动的射频对抗来自无效读取器的估计。MEAS由两个部分组成,一个估计干扰装置(EID)和两个精心设计的干扰消隐估计器(IBE)。随着标签的部署,EID会主动产生干扰信号,这会给无效或恶意的读取器带来足够大的估计误差。使用与EID共享的秘密干扰因子,有效的读取器可以通过两个ibe进行准确的估计。理论分析和仿真结果表明了MEAS的有效性。同时,MEAS还可以使用ibe保持较高的估计精度。
{"title":"Utilizing RF Interference to Enable Private Estimation in RFID Systems","authors":"Lei Yang, Jinsong Han, Yong Qi, Cheng Wang, Zhuo Li, Qingsong Yao, Ying Chen, Xiao Zhong","doi":"10.1109/ICPADS.2010.106","DOIUrl":"https://doi.org/10.1109/ICPADS.2010.106","url":null,"abstract":"Counting or estimating the number of tags is crucial for RFID system. Researchers have proposed several fast cardinality estimation schemes to estimate the quantity of a batch of tags within a short time frame. Existing estimation schemes scarcely consider the privacy issue. Without effective protection, the adversary can utilize the responding signals to estimate the number of tags as accurate as the valid reader. To address this issue, we propose a novel privacy-preserving estimation scheme, termed as MEAS, which provides an active RF countermeasure against the estimation from invalid readers. MEAS comprises of two components, an Estimation Interference Device (EID) and two well-designed Interference Blanking Estimators (IBE). EID is deployed with the tags to actively generate interfering signals, which introduce sufficiently large estimation errors to invalid or malicious readers. Using a secret interference factor shared with EID, a valid reader can perform accurate estimation via two IBEs. Our theoretical analysis and simulation results show the effectiveness of MEAS. Meanwhile, MEAS can also maintain a high estimation accuracy using IBEs.","PeriodicalId":365914,"journal":{"name":"2010 IEEE 16th International Conference on Parallel and Distributed Systems","volume":"120 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116367967","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Primal Dual Approach for Dynamic Bid Optimization 动态竞价优化的原始对偶方法
Pub Date : 2010-12-08 DOI: 10.1109/ICPADS.2010.75
Lingfei Yu, Kun She, Changyuan Yu
We study the dynamic bid optimization problem via a primal dual approach. In the case we have no information about the distribution of queries, we reconstruct the ln(U=L) + 1 competitive algorithm proposed in [ZCL08] through a systematic way and showed the intuition behind this algorithm. In the case of random permutation model, we showed that the learning technique used in [DH09] can give us a (1 ¡ O(²)) competitive algorithm for any small constant ² > 0 as long as the optimum is large enough.
利用原始对偶方法研究了动态竞价优化问题。在没有查询分布信息的情况下,我们通过系统的方式重构了[ZCL08]中提出的ln(U=L) + 1竞争算法,并展示了该算法背后的直觉。在随机排列模型的情况下,我们证明了[DH09]中使用的学习技术可以为我们提供一个(1±(²))竞争算法,对于任何小常数²> 0,只要最优值足够大。
{"title":"A Primal Dual Approach for Dynamic Bid Optimization","authors":"Lingfei Yu, Kun She, Changyuan Yu","doi":"10.1109/ICPADS.2010.75","DOIUrl":"https://doi.org/10.1109/ICPADS.2010.75","url":null,"abstract":"We study the dynamic bid optimization problem via a primal dual approach. In the case we have no information about the distribution of queries, we reconstruct the ln(U=L) + 1 competitive algorithm proposed in [ZCL08] through a systematic way and showed the intuition behind this algorithm. In the case of random permutation model, we showed that the learning technique used in [DH09] can give us a (1 ¡ O(²)) competitive algorithm for any small constant ² > 0 as long as the optimum is large enough.","PeriodicalId":365914,"journal":{"name":"2010 IEEE 16th International Conference on Parallel and Distributed Systems","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123932050","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
General vs. Interval Mappings for Streaming Applications 流应用程序的通用映射与间隔映射
Pub Date : 2010-12-08 DOI: 10.1109/ICPADS.2010.15
A. Benoit, Hinde-Lilia Bouziane, Y. Robert
This paper deals with the problem of mapping pipelined applications on heterogeneous platforms whose processors are subject to failures. We address a difficult bi-criteria problem, namely deciding which stages to replicate, and on which resources, in order to optimize the reliability of the schedule, while guaranteeing a minimal throughput. Previous work had addressed the complexity of interval mappings, where the application is partitioned into intervals of consecutive stages (which are then replicated and assigned to processors). In this paper we investigate general mappings, where stages may be partitioned without any constraint, thereby allowing a better usage of processors and communication network capabilities. The price to pay for general mappings is a dramatic increase in the problem complexity. We show that computing the period of a given general mapping is an NP-complete problem, and we provide polynomial bounds to determine a (conservative) approximated value. The bi-criteria mapping problem itself becomes NP-complete on homogeneous platforms, while it is polynomial with interval mappings. We design a set of efficient heuristics, which we compare with interval mapping strategies through extensive simulations.
本文研究了异构平台上的流水线应用程序的映射问题,这些平台的处理器可能会出现故障。我们解决了一个困难的双标准问题,即决定复制哪些阶段,以及在哪些资源上复制,以优化调度的可靠性,同时保证最小的吞吐量。以前的工作解决了间隔映射的复杂性,其中应用程序被划分为连续阶段的间隔(然后被复制并分配给处理器)。在本文中,我们研究一般映射,其中阶段可以不受任何约束地划分,从而允许更好地使用处理器和通信网络功能。为一般映射付出的代价是问题复杂性的急剧增加。我们证明了计算给定一般映射的周期是一个np完全问题,并且我们提供了多项式界来确定一个(保守的)近似值。双准则映射问题本身在齐次平台上是np完全的,而它是具有区间映射的多项式。我们设计了一套有效的启发式算法,并通过大量的仿真将其与区间映射策略进行了比较。
{"title":"General vs. Interval Mappings for Streaming Applications","authors":"A. Benoit, Hinde-Lilia Bouziane, Y. Robert","doi":"10.1109/ICPADS.2010.15","DOIUrl":"https://doi.org/10.1109/ICPADS.2010.15","url":null,"abstract":"This paper deals with the problem of mapping pipelined applications on heterogeneous platforms whose processors are subject to failures. We address a difficult bi-criteria problem, namely deciding which stages to replicate, and on which resources, in order to optimize the reliability of the schedule, while guaranteeing a minimal throughput. Previous work had addressed the complexity of interval mappings, where the application is partitioned into intervals of consecutive stages (which are then replicated and assigned to processors). In this paper we investigate general mappings, where stages may be partitioned without any constraint, thereby allowing a better usage of processors and communication network capabilities. The price to pay for general mappings is a dramatic increase in the problem complexity. We show that computing the period of a given general mapping is an NP-complete problem, and we provide polynomial bounds to determine a (conservative) approximated value. The bi-criteria mapping problem itself becomes NP-complete on homogeneous platforms, while it is polynomial with interval mappings. We design a set of efficient heuristics, which we compare with interval mapping strategies through extensive simulations.","PeriodicalId":365914,"journal":{"name":"2010 IEEE 16th International Conference on Parallel and Distributed Systems","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129735917","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Space Speedup and Its Relationship with Time Speedup 空间加速及其与时间加速的关系
Pub Date : 2010-12-08 DOI: 10.1109/ICPADS.2010.68
Yue Hu, W. Tong, Xiaoli Zhi, Zhi-xun Gong
In the optimizing work of parallel program, especially in the realm of massively parallel computing, the parallel computing time and space must be concurrently carefully considered to cut down the computing time as much as possible, because lots of poor parallel space strategies would impact negative effects on computing time. Although, sometimes we have no choice but to sacrifice the space for the time’s further diminishing. What relationship should the computing time and space to keep and how are they going on are two problems, which deciding our optimizing direction directly and must be clear in parallel optimizing. This paper proposes a space theory, named as space speedup, to denote the scalability of memory requirement, and discusses the relationship of time speedup and space speedup, through which the speedups’ guidance capacity in optimizing parallel codes are given.
在并行程序的优化工作中,特别是在大规模并行计算领域,必须同时考虑并行计算的时间和空间,以尽可能地减少计算时间,因为许多较差的并行空间策略会对计算时间产生负面影响。虽然,有时候我们不得不牺牲空间来换取时间的进一步减少。计算时间和空间应保持怎样的关系以及如何进行是直接决定优化方向的两个问题,在并行优化中必须弄清楚。本文提出了空间加速理论来表示存储需求的可扩展性,并讨论了时间加速与空间加速的关系,从而给出了加速对并行代码优化的引导能力。
{"title":"Space Speedup and Its Relationship with Time Speedup","authors":"Yue Hu, W. Tong, Xiaoli Zhi, Zhi-xun Gong","doi":"10.1109/ICPADS.2010.68","DOIUrl":"https://doi.org/10.1109/ICPADS.2010.68","url":null,"abstract":"In the optimizing work of parallel program, especially in the realm of massively parallel computing, the parallel computing time and space must be concurrently carefully considered to cut down the computing time as much as possible, because lots of poor parallel space strategies would impact negative effects on computing time. Although, sometimes we have no choice but to sacrifice the space for the time’s further diminishing. What relationship should the computing time and space to keep and how are they going on are two problems, which deciding our optimizing direction directly and must be clear in parallel optimizing. This paper proposes a space theory, named as space speedup, to denote the scalability of memory requirement, and discusses the relationship of time speedup and space speedup, through which the speedups’ guidance capacity in optimizing parallel codes are given.","PeriodicalId":365914,"journal":{"name":"2010 IEEE 16th International Conference on Parallel and Distributed Systems","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129274622","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Vapor: Virtual Machine Based Parallel Program Profiling Framework 蒸汽:基于虚拟机的并行程序分析框架
Pub Date : 2010-12-08 DOI: 10.1109/ICPADS.2010.59
Yusong Tan, Wei Chen, Q. Wu
It is hard to execute parallel program efficiently on man-core platform because we could not divide program into appropriate granularity executed simultaneously. Based on virtual machine and binary translation technologies the article proposes the vapor profiling framework that uses SBIRP instruction in-place replacement method to collect program’s run-time control flow and data flow information precisely. Moreover, it explains how to create control flow and data flow dependency graphs. Experiment results prove that vapor has better performance than traditional methods.
在人核平台上,由于不能将程序划分成适当的粒度同时执行,很难有效地执行并行程序。基于虚拟机和二进制翻译技术,提出了采用SBIRP指令就地替换方法精确采集程序运行时控制流和数据流信息的蒸汽剖面框架。此外,它还解释了如何创建控制流和数据流依赖关系图。实验结果表明,蒸汽法比传统方法具有更好的性能。
{"title":"Vapor: Virtual Machine Based Parallel Program Profiling Framework","authors":"Yusong Tan, Wei Chen, Q. Wu","doi":"10.1109/ICPADS.2010.59","DOIUrl":"https://doi.org/10.1109/ICPADS.2010.59","url":null,"abstract":"It is hard to execute parallel program efficiently on man-core platform because we could not divide program into appropriate granularity executed simultaneously. Based on virtual machine and binary translation technologies the article proposes the vapor profiling framework that uses SBIRP instruction in-place replacement method to collect program’s run-time control flow and data flow information precisely. Moreover, it explains how to create control flow and data flow dependency graphs. Experiment results prove that vapor has better performance than traditional methods.","PeriodicalId":365914,"journal":{"name":"2010 IEEE 16th International Conference on Parallel and Distributed Systems","volume":"96 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129819497","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Making Human Connectome Faster: GPU Acceleration of Brain Network Analysis 使人类连接组更快:GPU加速脑网络分析
Pub Date : 2010-12-08 DOI: 10.1109/ICPADS.2010.105
Di Wu, Tianji Wu, Yi Shan, Yu Wang, Yong He, Ningyi Xu, Huazhong Yang
The research on complex Brain Networks plays a vital role in understanding the connectivity patterns of the human brain and disease-related alterations. Recent studies have suggested a noninvasive way to model and analyze human brain networks by using multi-modal imaging and graph theoretical approaches. Both the construction and analysis of the Brain Networks require tremendous computation. As a result, most current studies of the Brain Networks are focused on a coarse scale based on Brain Regions. Networks on this scale usually consist around 100 nodes. The more accurate and meticulous voxel-base Brain Networks, on the other hand, may consist 20K to 100K nodes. In response to the difficulties of analyzing large-scale networks, we propose an acceleration framework for voxel-base Brain Network Analysis based on Graphics Processing Unit (GPU). Our GPU implementations of Brain Network construction and modularity achieve 24x and 80x speedup respectively, compared with single-core CPU. Our work makes the processing time affordable to analyze multiple large-scale Brain Networks.
复杂脑网络的研究对于理解人类大脑的连接模式和疾病相关的改变起着至关重要的作用。最近的研究提出了一种利用多模态成像和图理论方法对人脑网络进行建模和分析的无创方法。大脑网络的构建和分析都需要大量的计算。因此,目前大多数关于大脑网络的研究都集中在基于大脑区域的粗略尺度上。这种规模的网络通常由大约100个节点组成。另一方面,更精确和细致的基于体素的大脑网络可能包含20K到100K个节点。针对大规模网络分析的困难,提出了一种基于图形处理单元(GPU)的基于体素的脑网络分析加速框架。与单核CPU相比,我们的大脑网络构建和模块化GPU分别实现了24倍和80倍的加速。我们的工作使得处理时间可以负担得起分析多个大规模脑网络。
{"title":"Making Human Connectome Faster: GPU Acceleration of Brain Network Analysis","authors":"Di Wu, Tianji Wu, Yi Shan, Yu Wang, Yong He, Ningyi Xu, Huazhong Yang","doi":"10.1109/ICPADS.2010.105","DOIUrl":"https://doi.org/10.1109/ICPADS.2010.105","url":null,"abstract":"The research on complex Brain Networks plays a vital role in understanding the connectivity patterns of the human brain and disease-related alterations. Recent studies have suggested a noninvasive way to model and analyze human brain networks by using multi-modal imaging and graph theoretical approaches. Both the construction and analysis of the Brain Networks require tremendous computation. As a result, most current studies of the Brain Networks are focused on a coarse scale based on Brain Regions. Networks on this scale usually consist around 100 nodes. The more accurate and meticulous voxel-base Brain Networks, on the other hand, may consist 20K to 100K nodes. In response to the difficulties of analyzing large-scale networks, we propose an acceleration framework for voxel-base Brain Network Analysis based on Graphics Processing Unit (GPU). Our GPU implementations of Brain Network construction and modularity achieve 24x and 80x speedup respectively, compared with single-core CPU. Our work makes the processing time affordable to analyze multiple large-scale Brain Networks.","PeriodicalId":365914,"journal":{"name":"2010 IEEE 16th International Conference on Parallel and Distributed Systems","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127715986","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Smart: Service Model for Integrating Wireless Sensor Networks and the Internet 智能:无线传感器网络与互联网融合的服务模式
Pub Date : 2010-12-08 DOI: 10.1109/ICPADS.2010.92
Jeisa P. O. Domingues, A. Dâmaso, N. Rosa
Wireless sensor networks (WSNs) have received considerable attention in recent years as they have great potential for many distributed applications in different scenarios. Whatever the scenario, WSNs are actually connected to an external network, through which sensed information are passed to the Internet and control messages can reach the WSN. This paper presents Smart, a service model for integrating WSNs and the Internet at service level. Instead of integrating protocol stacks and/or mapping logical addresses, Smart allows the integration of Internet's and WSN's services by providing service interoperability. A communication infrastructure that implements the main components of Smart, along with a power consumption evaluation, is presented to validate the model.
近年来,无线传感器网络(WSNs)因其在不同场景下具有巨大的分布式应用潜力而受到广泛关注。无论哪种情况,WSN实际上都连接到外部网络,通过外部网络将感知到的信息传递到Internet,控制消息可以到达WSN。提出了一种在服务层面上集成无线传感器网络和Internet的服务模型Smart。Smart不是集成协议栈和/或映射逻辑地址,而是通过提供服务互操作性来集成Internet和WSN的服务。给出了实现Smart主要组件的通信基础设施,以及功耗评估来验证该模型。
{"title":"Smart: Service Model for Integrating Wireless Sensor Networks and the Internet","authors":"Jeisa P. O. Domingues, A. Dâmaso, N. Rosa","doi":"10.1109/ICPADS.2010.92","DOIUrl":"https://doi.org/10.1109/ICPADS.2010.92","url":null,"abstract":"Wireless sensor networks (WSNs) have received considerable attention in recent years as they have great potential for many distributed applications in different scenarios. Whatever the scenario, WSNs are actually connected to an external network, through which sensed information are passed to the Internet and control messages can reach the WSN. This paper presents Smart, a service model for integrating WSNs and the Internet at service level. Instead of integrating protocol stacks and/or mapping logical addresses, Smart allows the integration of Internet's and WSN's services by providing service interoperability. A communication infrastructure that implements the main components of Smart, along with a power consumption evaluation, is presented to validate the model.","PeriodicalId":365914,"journal":{"name":"2010 IEEE 16th International Conference on Parallel and Distributed Systems","volume":"112 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128049056","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
2F: A Special Cache for Mapping Table of Page-Level Flash Translation Layer 一种用于页面级Flash翻译层映射表的特殊缓存
Pub Date : 2010-12-08 DOI: 10.1109/ICPADS.2010.60
Zhiguang Chen, Nong Xiao, Fang Liu, Yimo Du
The development of flash memory drives flash based SSDs to enter into enterprise-scale storage systems. As the kernel of SSD, flash translation layer (FTL) attracts many attentions. Generally, there are two types of FTLs according to the granularity of address mapping: block-level and page-level mapping FTLs. We focus on the latter one. Typically, page-level mapping scheme must employ a cache to alleviate the memory pressure introduced by the big mapping table. We argue that classic cache replacement policies aren’t competent for the page table cache of FTLs. The major contribution of this work is to design a dedicated cache replacement policy called Two Filters (abbreviated as 2F) for page-level mapping FTLs. 2F aims at two goals. The first is higher hit ratio as all the replacement policies pursue. As 2F not only protects frequently accessed pages, but also protects sequentially accessed pages at little cost, it does achieve a higher hit ratio. The second goal is to distinguish hot pages from the cold. This goal is special for page table of FTLs. If hot and cold pages are directed to separate blocks, garbage collection will be more efficient. In order to achieve this goal, 2F employs two filters. One is used for containing sequentially accessed pages. Another is used for selecting hot pages. Trace driven simulations present that 2F outperforms classic replacement policies in both hit ratio and data classification.
闪存驱动器的发展使基于闪存的ssd进入企业级存储系统。作为固态硬盘的核心,闪存转换层(FTL)备受关注。根据地址映射的粒度不同,一般分为块级映射FTLs和页级映射FTLs。我们关注的是后者。通常,页级映射方案必须使用缓存来减轻大映射表带来的内存压力。我们认为,经典的缓存替换策略不适用于ftl的页表缓存。这项工作的主要贡献是为页级映射ftl设计了一个专用的缓存替换策略,称为Two Filters(缩写为2F)。2F有两个目标。首先是所有替代政策都追求更高的命中率。由于2F不仅可以保护频繁访问的页面,而且可以以很小的成本保护顺序访问的页面,因此可以实现更高的命中率。第二个目标是区分热页和冷页。这个目标特别适用于ftl的页表。如果将热页和冷页定向到单独的块,那么垃圾收集将更有效。为了实现这一目标,2F使用了两个过滤器。一个用于包含顺序访问的页面。另一个用于选择热门页面。跟踪驱动模拟表明,2F在命中率和数据分类方面都优于经典替换策略。
{"title":"2F: A Special Cache for Mapping Table of Page-Level Flash Translation Layer","authors":"Zhiguang Chen, Nong Xiao, Fang Liu, Yimo Du","doi":"10.1109/ICPADS.2010.60","DOIUrl":"https://doi.org/10.1109/ICPADS.2010.60","url":null,"abstract":"The development of flash memory drives flash based SSDs to enter into enterprise-scale storage systems. As the kernel of SSD, flash translation layer (FTL) attracts many attentions. Generally, there are two types of FTLs according to the granularity of address mapping: block-level and page-level mapping FTLs. We focus on the latter one. Typically, page-level mapping scheme must employ a cache to alleviate the memory pressure introduced by the big mapping table. We argue that classic cache replacement policies aren’t competent for the page table cache of FTLs. The major contribution of this work is to design a dedicated cache replacement policy called Two Filters (abbreviated as 2F) for page-level mapping FTLs. 2F aims at two goals. The first is higher hit ratio as all the replacement policies pursue. As 2F not only protects frequently accessed pages, but also protects sequentially accessed pages at little cost, it does achieve a higher hit ratio. The second goal is to distinguish hot pages from the cold. This goal is special for page table of FTLs. If hot and cold pages are directed to separate blocks, garbage collection will be more efficient. In order to achieve this goal, 2F employs two filters. One is used for containing sequentially accessed pages. Another is used for selecting hot pages. Trace driven simulations present that 2F outperforms classic replacement policies in both hit ratio and data classification.","PeriodicalId":365914,"journal":{"name":"2010 IEEE 16th International Conference on Parallel and Distributed Systems","volume":"100 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121395978","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
2010 IEEE 16th International Conference on Parallel and Distributed Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1