首页 > 最新文献

Proceedings of the Third ACM International Workshop on Edge Systems, Analytics and Networking最新文献

英文 中文
Aspect-oriented language for reactive distributed applications at the edge 面向方面的语言,用于边缘的响应式分布式应用
I. Kuraj, Armando Solar-Lezama
This paper presents EdgeC, a new language for programming reactive distributed applications. It enables separation of concerns between expressing behavior and controlling distributed aspects, inspired by aspect-oriented language design. In EdgeC, developers express functionality with sequential behaviors, and data allocation, reactivity, consistency, and underlying network with orthogonal specifications. Through such separation, EdgeC allows developers to change functionality and control the shape of resulting distributed behaviors without cross-cutting code, simplifying deployment to the edge. Developers can reason about and test their applications as sequential executions, whilst EdgeC automatically synthesizes low-level distributed code. It handles, with the help of the EdgeC run-time, allocation, communication, concurrency, and coordination, across the specified, potentially non-uniform, network model. We introduce the main features of EdgeC, present the new compiler design, its prototype implementation, the resulting performance, and discuss the potential of the approach for simplifying development of reactive applications over nonuniform networks and achieving performance gains, compared to existing approaches.
本文介绍了一种新的响应式分布式应用程序编程语言EdgeC。受面向方面语言设计的启发,它实现了表达行为和控制分布式方面之间的关注点分离。在EdgeC中,开发人员用顺序行为表达功能,用正交规范表达数据分配、反应性、一致性和底层网络。通过这样的分离,EdgeC允许开发人员在没有横切代码的情况下更改功能和控制生成的分布式行为的形状,从而简化了对边缘的部署。开发人员可以按照顺序执行对应用程序进行推理和测试,而EdgeC则自动合成低级分布式代码。在EdgeC运行时的帮助下,它跨指定的、可能不统一的网络模型处理分配、通信、并发性和协调。我们介绍了EdgeC的主要特性,介绍了新的编译器设计,它的原型实现,产生的性能,并讨论了与现有方法相比,该方法在简化非统一网络上响应式应用程序的开发和实现性能提升方面的潜力。
{"title":"Aspect-oriented language for reactive distributed applications at the edge","authors":"I. Kuraj, Armando Solar-Lezama","doi":"10.1145/3378679.3394531","DOIUrl":"https://doi.org/10.1145/3378679.3394531","url":null,"abstract":"This paper presents EdgeC, a new language for programming reactive distributed applications. It enables separation of concerns between expressing behavior and controlling distributed aspects, inspired by aspect-oriented language design. In EdgeC, developers express functionality with sequential behaviors, and data allocation, reactivity, consistency, and underlying network with orthogonal specifications. Through such separation, EdgeC allows developers to change functionality and control the shape of resulting distributed behaviors without cross-cutting code, simplifying deployment to the edge. Developers can reason about and test their applications as sequential executions, whilst EdgeC automatically synthesizes low-level distributed code. It handles, with the help of the EdgeC run-time, allocation, communication, concurrency, and coordination, across the specified, potentially non-uniform, network model. We introduce the main features of EdgeC, present the new compiler design, its prototype implementation, the resulting performance, and discuss the potential of the approach for simplifying development of reactive applications over nonuniform networks and achieving performance gains, compared to existing approaches.","PeriodicalId":268360,"journal":{"name":"Proceedings of the Third ACM International Workshop on Edge Systems, Analytics and Networking","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115206263","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
LDP-Fed: federated learning with local differential privacy LDP-Fed:具有局部差分隐私的联邦学习
Stacey Truex, Ling Liu, Ka-Ho Chow, M. E. Gursoy, Wenqi Wei
This paper presents LDP-Fed, a novel federated learning system with a formal privacy guarantee using local differential privacy (LDP). Existing LDP protocols are developed primarily to ensure data privacy in the collection of single numerical or categorical values, such as click count in Web access logs. However, in federated learning model parameter updates are collected iteratively from each participant and consist of high dimensional, continuous values with high precision (10s of digits after the decimal point), making existing LDP protocols inapplicable. To address this challenge in LDP-Fed, we design and develop two novel approaches. First, LDP-Fed's LDP Module provides a formal differential privacy guarantee for the repeated collection of model training parameters in the federated training of large-scale neural networks over multiple individual participants' private datasets. Second, LDP-Fed implements a suite of selection and filtering techniques for perturbing and sharing select parameter updates with the parameter server. We validate our system deployed with a condensed LDP protocol in training deep neural networks on public data. We compare this version of LDP-Fed, coined CLDP-Fed, with other state-of-the-art approaches with respect to model accuracy, privacy preservation, and system capabilities.
本文提出了一种利用局部差分隐私(LDP)实现形式隐私保证的新型联邦学习系统LDP- fed。现有的LDP协议主要是为了确保单个数值或分类值集合中的数据隐私,例如Web访问日志中的点击计数。然而,在联邦学习模型中,参数更新是从每个参与者迭代地收集的,并且由高精度的高维连续值(小数点后的10位数)组成,使得现有的LDP协议不适用。为了解决LDP-Fed中的这一挑战,我们设计并开发了两种新颖的方法。首先,LDP- fed的LDP模块为大规模神经网络在多个个体参与者私有数据集上的联合训练中模型训练参数的重复收集提供了正式的差分隐私保证。其次,LDP-Fed实现了一套选择和过滤技术,用于干扰和与参数服务器共享选择参数更新。我们用压缩LDP协议在公共数据上训练深度神经网络来验证我们的系统。我们将这个版本的LDP-Fed(被称为CLDP-Fed)与其他最先进的方法在模型准确性、隐私保护和系统功能方面进行了比较。
{"title":"LDP-Fed: federated learning with local differential privacy","authors":"Stacey Truex, Ling Liu, Ka-Ho Chow, M. E. Gursoy, Wenqi Wei","doi":"10.1145/3378679.3394533","DOIUrl":"https://doi.org/10.1145/3378679.3394533","url":null,"abstract":"This paper presents LDP-Fed, a novel federated learning system with a formal privacy guarantee using local differential privacy (LDP). Existing LDP protocols are developed primarily to ensure data privacy in the collection of single numerical or categorical values, such as click count in Web access logs. However, in federated learning model parameter updates are collected iteratively from each participant and consist of high dimensional, continuous values with high precision (10s of digits after the decimal point), making existing LDP protocols inapplicable. To address this challenge in LDP-Fed, we design and develop two novel approaches. First, LDP-Fed's LDP Module provides a formal differential privacy guarantee for the repeated collection of model training parameters in the federated training of large-scale neural networks over multiple individual participants' private datasets. Second, LDP-Fed implements a suite of selection and filtering techniques for perturbing and sharing select parameter updates with the parameter server. We validate our system deployed with a condensed LDP protocol in training deep neural networks on public data. We compare this version of LDP-Fed, coined CLDP-Fed, with other state-of-the-art approaches with respect to model accuracy, privacy preservation, and system capabilities.","PeriodicalId":268360,"journal":{"name":"Proceedings of the Third ACM International Workshop on Edge Systems, Analytics and Networking","volume":"311 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132567778","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 205
CoLearn CoLearn
Angelo Feraudo, Poonam Yadav, Vadim Safronov, Diana Andreea Popescu, R. Mortier, Shiqiang Wang, P. Bellavista, J. Crowcroft
Edge computing and Federated Learning (FL) can work in tandem to address issues related to privacy and collaborative distributed learning in untrusted IoT environments. However, deployment of FL in resource-constrained IoT devices faces challenges including asynchronous participation of such devices in training, and the need to prevent malicious devices from participating. To address these challenges we present CoLearn, which build on the open-source Manufacturer Usage Description (MUD) implementation osMUD and the FL framework PySyft. We deploy CoLearn on resource-constrained devices in a lab environment to demonstrate (i) an asynchronous participation mechanism for IoT devices in machine learning model training using a publish/subscribe architecture, (ii) a mechanism for reducing the attack surface in FL architecture by allowing only IoT MUD-compliant devices to participate in the training phases, and (iii) a trade-off between communication bandwidth usage, training time and device temperature (thermal fatigue).
{"title":"CoLearn","authors":"Angelo Feraudo, Poonam Yadav, Vadim Safronov, Diana Andreea Popescu, R. Mortier, Shiqiang Wang, P. Bellavista, J. Crowcroft","doi":"10.1145/3378679.3394528","DOIUrl":"https://doi.org/10.1145/3378679.3394528","url":null,"abstract":"Edge computing and Federated Learning (FL) can work in tandem to address issues related to privacy and collaborative distributed learning in untrusted IoT environments. However, deployment of FL in resource-constrained IoT devices faces challenges including asynchronous participation of such devices in training, and the need to prevent malicious devices from participating. To address these challenges we present CoLearn, which build on the open-source Manufacturer Usage Description (MUD) implementation osMUD and the FL framework PySyft. We deploy CoLearn on resource-constrained devices in a lab environment to demonstrate (i) an asynchronous participation mechanism for IoT devices in machine learning model training using a publish/subscribe architecture, (ii) a mechanism for reducing the attack surface in FL architecture by allowing only IoT MUD-compliant devices to participate in the training phases, and (iii) a trade-off between communication bandwidth usage, training time and device temperature (thermal fatigue).","PeriodicalId":268360,"journal":{"name":"Proceedings of the Third ACM International Workshop on Edge Systems, Analytics and Networking","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121222417","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 39
An enclave assisted snapshot-based kernel integrity monitor 一个enclave辅助的基于快照的内核完整性监视器
Dimitris Deyannis, Dimitris Karnikis, G. Vasiliadis, S. Ioannidis
The integrity of operating system (OS) kernels is of paramount importance in order to ensure the secure operation of user-level processes and services as well as the benign behavior of the entire system. Attackers aim to exploit a system's kernel since compromising it provides more flexibility for malicious operations compared to compromising a user-level process. Acquiring access to the OS kernel enables malicious parties to manipulate process execution, control the file system and the peripheral devices and obtain securityand privacy-critical data. One of the most effective countermeasures against rootkits are kernel integrity monitors, implemented in software (often assisted by a hypervisor) or external hardware, aiming to detect threats by scanning the kernel's state. However, modern rootkits are able to hide their presence and prevent detection from such mechanisms either by identifying and disabling the monitors or by performing transient attacks. In this paper we present SGX-Mon, an external kernel integrity monitor that verifies the operating system's kernel integrity using a very small TCB while it does not require any OS modifications or external hardware. SGX-Mon is a snapshot-based monitor, residing in the user space, and utilizes the trusted execution environment offered by Intel SGX enclaves in order to avoid detection from rootkits and prevent attackers from tampering its execution and operation-critical data. Our system is able to perform scanning, analysis and verification of arbitrary kernel memory pages and memory regions and ensure their integrity. The monitored locations can be specified by the user and can contain critical kernel code and data. SGX-Mon scans the system periodically and compares the contents of critical memory regions against their known benign values. Our experimental results show that SGX-Mon is able to achieve 100% accuracy while scanning up to 6,000 distinct kernel memory locations.
为了确保用户级进程和服务的安全运行以及整个系统的良好行为,操作系统内核的完整性至关重要。攻击者的目标是利用系统的内核,因为与破坏用户级进程相比,破坏系统内核为恶意操作提供了更大的灵活性。获得对操作系统内核的访问使恶意方能够操纵进程执行,控制文件系统和外围设备,并获得安全和隐私关键数据。针对rootkit的最有效对策之一是内核完整性监视器,它在软件(通常由管理程序辅助)或外部硬件中实现,旨在通过扫描内核状态来检测威胁。然而,现代rootkit能够隐藏它们的存在,并通过识别和禁用监视器或执行瞬态攻击来阻止此类机制的检测。在本文中,我们介绍了SGX-Mon,这是一个外部内核完整性监视器,它使用非常小的TCB来验证操作系统的内核完整性,同时不需要任何操作系统修改或外部硬件。SGX- mon是一个基于快照的监视器,驻留在用户空间中,并利用英特尔SGX enclaves提供的可信执行环境,以避免来自rootkit的检测,并防止攻击者篡改其执行和操作关键数据。我们的系统能够对任意内核内存页面和内存区域进行扫描、分析和验证,并确保其完整性。被监视的位置可以由用户指定,并且可以包含关键的内核代码和数据。SGX-Mon定期扫描系统,并将关键内存区域的内容与其已知的良性值进行比较。我们的实验结果表明,SGX-Mon能够在扫描多达6,000个不同的内核内存位置时达到100%的准确性。
{"title":"An enclave assisted snapshot-based kernel integrity monitor","authors":"Dimitris Deyannis, Dimitris Karnikis, G. Vasiliadis, S. Ioannidis","doi":"10.1145/3378679.3394539","DOIUrl":"https://doi.org/10.1145/3378679.3394539","url":null,"abstract":"The integrity of operating system (OS) kernels is of paramount importance in order to ensure the secure operation of user-level processes and services as well as the benign behavior of the entire system. Attackers aim to exploit a system's kernel since compromising it provides more flexibility for malicious operations compared to compromising a user-level process. Acquiring access to the OS kernel enables malicious parties to manipulate process execution, control the file system and the peripheral devices and obtain securityand privacy-critical data. One of the most effective countermeasures against rootkits are kernel integrity monitors, implemented in software (often assisted by a hypervisor) or external hardware, aiming to detect threats by scanning the kernel's state. However, modern rootkits are able to hide their presence and prevent detection from such mechanisms either by identifying and disabling the monitors or by performing transient attacks. In this paper we present SGX-Mon, an external kernel integrity monitor that verifies the operating system's kernel integrity using a very small TCB while it does not require any OS modifications or external hardware. SGX-Mon is a snapshot-based monitor, residing in the user space, and utilizes the trusted execution environment offered by Intel SGX enclaves in order to avoid detection from rootkits and prevent attackers from tampering its execution and operation-critical data. Our system is able to perform scanning, analysis and verification of arbitrary kernel memory pages and memory regions and ensure their integrity. The monitored locations can be specified by the user and can contain critical kernel code and data. SGX-Mon scans the system periodically and compares the contents of critical memory regions against their known benign values. Our experimental results show that SGX-Mon is able to achieve 100% accuracy while scanning up to 6,000 distinct kernel memory locations.","PeriodicalId":268360,"journal":{"name":"Proceedings of the Third ACM International Workshop on Edge Systems, Analytics and Networking","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129102683","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
On the impact of clustering for IoT analytics and message broker placement across cloud and edge 关于集群对物联网分析和跨云和边缘的消息代理放置的影响
Daniel Happ, S. Bayhan
With edge computing emerging as a promising solution to cope with the challenges of Internet of Things (IoT) systems, there is an increasing need to automate the deployment of large-scale applications along with the publish/subscribe brokers they communicate over. Such a placement must adjust to the resource requirements of both applications and brokers in the heterogeneous environment of edge, fog, and cloud. In contrast to prior work focusing only on the placement of applications, this paper addresses the problem of jointly placing IoT applications and the pub/sub brokers on a set of network nodes, considering an application provider who aims at minimizing total end-to-end delays of all its subscribers. More specifically, we devise two heuristics for joint deployment of brokers and applications and analyze their performance in comparison to the current cloud-based IoT solutions wherein both the IoT applications and the brokers are located solely in the cloud. As an application provider should consider not only the location of the application users but also how they are distributed across different network components, we use von Mises distributions to model the degree of clustering of the users of an IoT application. Our simulations show that superior performance of our heuristics in comparison to cloud-based IoT operation is most pronounced under a high degree of clustering. When users of an IoT application are in close network proximity of the IoT sensors, cloud-based IoT unnecessarily introduces latency to move the data from the edge to the cloud and vice versa while processing could be performed at the edge or the fog layers.
随着边缘计算成为应对物联网(IoT)系统挑战的一种有前途的解决方案,对大规模应用程序及其通信的发布/订阅代理的自动化部署的需求越来越大。这样的放置必须适应边缘、雾和云等异构环境中应用程序和代理的资源需求。与之前的工作只关注应用程序的放置相反,本文解决了在一组网络节点上联合放置物联网应用程序和pub/sub代理的问题,考虑到应用程序提供商的目标是最小化其所有订户的总端到端延迟。更具体地说,我们为代理和应用程序的联合部署设计了两种启发式方法,并与当前基于云的物联网解决方案(物联网应用程序和代理都单独位于云中)进行了比较,分析了它们的性能。由于应用程序提供商不仅要考虑应用程序用户的位置,还要考虑他们如何分布在不同的网络组件中,因此我们使用von Mises分布来模拟物联网应用程序用户的集群程度。我们的模拟表明,与基于云的物联网操作相比,我们的启发式算法的优越性能在高度集群下最为明显。当物联网应用程序的用户靠近物联网传感器时,基于云的物联网不必要地引入延迟,将数据从边缘移动到云,反之亦然,而处理可以在边缘或雾层执行。
{"title":"On the impact of clustering for IoT analytics and message broker placement across cloud and edge","authors":"Daniel Happ, S. Bayhan","doi":"10.1145/3378679.3394538","DOIUrl":"https://doi.org/10.1145/3378679.3394538","url":null,"abstract":"With edge computing emerging as a promising solution to cope with the challenges of Internet of Things (IoT) systems, there is an increasing need to automate the deployment of large-scale applications along with the publish/subscribe brokers they communicate over. Such a placement must adjust to the resource requirements of both applications and brokers in the heterogeneous environment of edge, fog, and cloud. In contrast to prior work focusing only on the placement of applications, this paper addresses the problem of jointly placing IoT applications and the pub/sub brokers on a set of network nodes, considering an application provider who aims at minimizing total end-to-end delays of all its subscribers. More specifically, we devise two heuristics for joint deployment of brokers and applications and analyze their performance in comparison to the current cloud-based IoT solutions wherein both the IoT applications and the brokers are located solely in the cloud. As an application provider should consider not only the location of the application users but also how they are distributed across different network components, we use von Mises distributions to model the degree of clustering of the users of an IoT application. Our simulations show that superior performance of our heuristics in comparison to cloud-based IoT operation is most pronounced under a high degree of clustering. When users of an IoT application are in close network proximity of the IoT sensors, cloud-based IoT unnecessarily introduces latency to move the data from the edge to the cloud and vice versa while processing could be performed at the edge or the fog layers.","PeriodicalId":268360,"journal":{"name":"Proceedings of the Third ACM International Workshop on Edge Systems, Analytics and Networking","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116822937","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
PAIGE 佩奇
Yilei Liang, Daniel O'keeffe, Nishanth R. Sastry
Intelligent Personal Assistants (IPAs) such as Apple's Siri, Google Now, and Amazon Alexa are becoming an increasingly important class of web application. In contrast to previous keyword-oriented search applications, IPAs support a rich query interface that allows user interaction through images, audio, and natural language queries. However, modern IPAs rely heavily on compute-intensive machine-learning inference. To achieve acceptable performance, ML-driven IPAs increasingly depend on specialized hardware accelerators (e.g. GPUs, FPGAs or TPUs), increasing costs for IPA service providers. For end-users, IPAs also present considerable privacy risks given the sensitive nature of the data they capture. We present PAIGE, a hybrid edge-cloud architecture for privacy-preserving Intelligent Personal Assistants. PAIGE's design is founded on the assumption that recent advances in low-cost hardware for machine-learning inference offer an opportunity to offload compute-intensive IPA ML tasks to the network edge. To allow privacy-preserving access to large IPA databases for less compute-intensive pre-processed queries, PAIGE leverages trusted execution environments at the server side. PAIGE's hybrid design allows privacy-preserving hardware acceleration of compute-intensive tasks, while avoiding the need to move potentially large IPA question-answering databases to the edge. As a step towards realising PAIGE, we present a first systematic performance evaluation of existing edge accelerator hardware platforms for a subset of IPA workloads, and show they offer a competitive alternative to existing data-center alternatives.
{"title":"PAIGE","authors":"Yilei Liang, Daniel O'keeffe, Nishanth R. Sastry","doi":"10.1145/3378679.3394536","DOIUrl":"https://doi.org/10.1145/3378679.3394536","url":null,"abstract":"Intelligent Personal Assistants (IPAs) such as Apple's Siri, Google Now, and Amazon Alexa are becoming an increasingly important class of web application. In contrast to previous keyword-oriented search applications, IPAs support a rich query interface that allows user interaction through images, audio, and natural language queries. However, modern IPAs rely heavily on compute-intensive machine-learning inference. To achieve acceptable performance, ML-driven IPAs increasingly depend on specialized hardware accelerators (e.g. GPUs, FPGAs or TPUs), increasing costs for IPA service providers. For end-users, IPAs also present considerable privacy risks given the sensitive nature of the data they capture. We present PAIGE, a hybrid edge-cloud architecture for privacy-preserving Intelligent Personal Assistants. PAIGE's design is founded on the assumption that recent advances in low-cost hardware for machine-learning inference offer an opportunity to offload compute-intensive IPA ML tasks to the network edge. To allow privacy-preserving access to large IPA databases for less compute-intensive pre-processed queries, PAIGE leverages trusted execution environments at the server side. PAIGE's hybrid design allows privacy-preserving hardware acceleration of compute-intensive tasks, while avoiding the need to move potentially large IPA question-answering databases to the edge. As a step towards realising PAIGE, we present a first systematic performance evaluation of existing edge accelerator hardware platforms for a subset of IPA workloads, and show they offer a competitive alternative to existing data-center alternatives.","PeriodicalId":268360,"journal":{"name":"Proceedings of the Third ACM International Workshop on Edge Systems, Analytics and Networking","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122179616","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Towards federated unsupervised representation learning 迈向联邦无监督表示学习
Bram van Berlo, Aaqib Saeed, T. Ozcelebi
Making deep learning models efficient at inferring nowadays requires training with an extensive number of labeled data that are gathered in a centralized system. However, gathering labeled data is an expensive and time-consuming process, centralized systems cannot aggregate an ever-increasing amount of data and aggregating user data is raising privacy concerns. Federated learning solves data volume and privacy issues by leaving user data on devices, but is limited to use cases where labeled data can be generated from user interaction. Unsupervised representation learning reduces the amount of labeled data required for model training, but previous work is limited to centralized systems. This work introduces federated unsupervised representation learning, a novel software architecture that uses unsupervised representation learning to pre-train deep neural networks using unlabeled data in a federated setting. Pre-trained networks can be used to extract discriminative features. The features help learn a down-stream task of interest with a reduced amount of labeled data. Based on representation performance experiments with human activity detection it is recommended to pre-train with unlabeled data originating from more users performing a bigger set of activities compared to data used with the down-stream task of interest. As a result, competitive or superior performance compared to supervised deep learning is achieved.
如今,要使深度学习模型在推理方面高效,需要使用集中系统中收集的大量标记数据进行训练。然而,收集标记数据是一个昂贵且耗时的过程,集中式系统无法聚合不断增加的数据量,并且聚合用户数据会引起隐私问题。联邦学习通过将用户数据留在设备上来解决数据量和隐私问题,但仅限于可以从用户交互生成标记数据的用例。无监督表示学习减少了模型训练所需的标记数据量,但以前的工作仅限于集中式系统。这项工作引入了联邦无监督表示学习,这是一种新的软件架构,它使用无监督表示学习在联邦设置中使用未标记数据预训练深度神经网络。预训练的网络可以用来提取判别特征。这些特性有助于通过减少标记数据量来学习感兴趣的下游任务。基于人类活动检测的表征性能实验,建议使用来自更多用户执行更大活动集的未标记数据进行预训练,而不是使用下游感兴趣任务使用的数据。因此,与监督深度学习相比,实现了具有竞争力或更好的性能。
{"title":"Towards federated unsupervised representation learning","authors":"Bram van Berlo, Aaqib Saeed, T. Ozcelebi","doi":"10.1145/3378679.3394530","DOIUrl":"https://doi.org/10.1145/3378679.3394530","url":null,"abstract":"Making deep learning models efficient at inferring nowadays requires training with an extensive number of labeled data that are gathered in a centralized system. However, gathering labeled data is an expensive and time-consuming process, centralized systems cannot aggregate an ever-increasing amount of data and aggregating user data is raising privacy concerns. Federated learning solves data volume and privacy issues by leaving user data on devices, but is limited to use cases where labeled data can be generated from user interaction. Unsupervised representation learning reduces the amount of labeled data required for model training, but previous work is limited to centralized systems. This work introduces federated unsupervised representation learning, a novel software architecture that uses unsupervised representation learning to pre-train deep neural networks using unlabeled data in a federated setting. Pre-trained networks can be used to extract discriminative features. The features help learn a down-stream task of interest with a reduced amount of labeled data. Based on representation performance experiments with human activity detection it is recommended to pre-train with unlabeled data originating from more users performing a bigger set of activities compared to data used with the down-stream task of interest. As a result, competitive or superior performance compared to supervised deep learning is achieved.","PeriodicalId":268360,"journal":{"name":"Proceedings of the Third ACM International Workshop on Edge Systems, Analytics and Networking","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129907284","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 55
The serverkernel operating system serverkernel操作系统
Jon Larrea, A. Barbalace
With the idea of exploiting all the computational resources that an IoT environment with multiple interconnected devices offers, serverkernel is presented as a new operating system architecture that blends ideas from distributed operating systems, Unikernel, and LWK. These concepts are mixed with a server in which a user can remotely offload computations and get the result. This single space-address operating system (OS) can be interpreted as a bare-metal OS in which only drivers for CPU, network, and accelerators are required in order to provide service. To demonstrate the advantages of serverkernel, jonOS, an open-source C implementation of this architecture for Raspberry Pi, is provided. Compared with commercial architectures used in IoT devices, serverkernel achieves an improvement ratio of 1.5 in CPU time, 2.5 in real-time, and around 9 times better in network speed.
考虑到利用具有多个互联设备的物联网环境提供的所有计算资源,serverkernel作为一种新的操作系统架构提出,它融合了分布式操作系统、Unikernel和LWK的思想。这些概念与服务器混合在一起,用户可以在其中远程卸载计算并获得结果。这种单空间地址操作系统(OS)可以解释为裸机操作系统,其中只需要CPU、网络和加速器的驱动程序就可以提供服务。为了演示服务器内核的优点,提供了用于树莓派的这种架构的开源C实现jonOS。与物联网设备中使用的商用架构相比,serverkernel的CPU时间提升了1.5倍,实时性提升了2.5倍,网速提升了约9倍。
{"title":"The serverkernel operating system","authors":"Jon Larrea, A. Barbalace","doi":"10.1145/3378679.3394537","DOIUrl":"https://doi.org/10.1145/3378679.3394537","url":null,"abstract":"With the idea of exploiting all the computational resources that an IoT environment with multiple interconnected devices offers, serverkernel is presented as a new operating system architecture that blends ideas from distributed operating systems, Unikernel, and LWK. These concepts are mixed with a server in which a user can remotely offload computations and get the result. This single space-address operating system (OS) can be interpreted as a bare-metal OS in which only drivers for CPU, network, and accelerators are required in order to provide service. To demonstrate the advantages of serverkernel, jonOS, an open-source C implementation of this architecture for Raspberry Pi, is provided. Compared with commercial architectures used in IoT devices, serverkernel achieves an improvement ratio of 1.5 in CPU time, 2.5 in real-time, and around 9 times better in network speed.","PeriodicalId":268360,"journal":{"name":"Proceedings of the Third ACM International Workshop on Edge Systems, Analytics and Networking","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115379179","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Quantifying the latency benefits of near-edge and in-network FPGA acceleration 量化近边和网络内FPGA加速的延迟效益
Ryan A. Cooke, Suhaib A. Fahmy
Transmitting data to cloud datacenters in distributed IoT applications introduces significant communication latency, but is often the only feasible solution when source nodes are computationally limited. To address latency concerns, Cloudlets, in-network computing, and more capable edge nodes are all being explored as a way of moving processing capability towards the edge of the network. Hardware acceleration using Field programmable gate arrays (FPGAs) is also seeing increased interest due to reduced computation time and improved efficiency. This paper evaluates the the implications of these offloading approaches using a case study neural network based image classification application, quantifying both the computation and communication latency resulting from different platform choices. We demonstrate that emerging in-network accelerator approaches offer much improved and predictable performance as well as better scaling to support multiple data sources.
在分布式物联网应用程序中,将数据传输到云数据中心会引入显著的通信延迟,但当源节点计算有限时,这通常是唯一可行的解决方案。为了解决延迟问题,人们正在探索Cloudlets、网络内计算和更强大的边缘节点,以将处理能力转移到网络边缘。由于减少了计算时间和提高了效率,使用现场可编程门阵列(fpga)的硬件加速也受到越来越多的关注。本文以一个基于神经网络的图像分类应用为例,评估了这些卸载方法的影响,量化了不同平台选择导致的计算和通信延迟。我们证明了新兴的网络内加速器方法提供了大大改进和可预测的性能,以及更好的可扩展性以支持多个数据源。
{"title":"Quantifying the latency benefits of near-edge and in-network FPGA acceleration","authors":"Ryan A. Cooke, Suhaib A. Fahmy","doi":"10.1145/3378679.3394534","DOIUrl":"https://doi.org/10.1145/3378679.3394534","url":null,"abstract":"Transmitting data to cloud datacenters in distributed IoT applications introduces significant communication latency, but is often the only feasible solution when source nodes are computationally limited. To address latency concerns, Cloudlets, in-network computing, and more capable edge nodes are all being explored as a way of moving processing capability towards the edge of the network. Hardware acceleration using Field programmable gate arrays (FPGAs) is also seeing increased interest due to reduced computation time and improved efficiency. This paper evaluates the the implications of these offloading approaches using a case study neural network based image classification application, quantifying both the computation and communication latency resulting from different platform choices. We demonstrate that emerging in-network accelerator approaches offer much improved and predictable performance as well as better scaling to support multiple data sources.","PeriodicalId":268360,"journal":{"name":"Proceedings of the Third ACM International Workshop on Edge Systems, Analytics and Networking","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124635965","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Privacy-preserving activity and health monitoring on databox 数据箱上的隐私保护活动和运行状况监控
Yuchen Zhao, H. Haddadi, Severin Skillman, Shirin Enshaeifar, P. Barnaghi
Activity recognition using deep learning and sensor data can help monitor activities and health conditions of people who need assistance in their daily lives. Deep Neural Network (DNN) models to infer the activities require data collected by in-home sensory devices. These data are often sent to a centralised cloud to be used for training the model. Centralising the data introduces privacy risks. The collected data contain sensitive information about the subjects. The cloud-based approach increases the risk that the data be stored and reused for other purposes without the owner's control. We propose a system that uses edge devices to implement activity and health monitoring locally and applies federated learning to facilitate the training process. The devices use the Databox platform to manage sensor data collected in people's homes, conduct activity recognition locally, and collaboratively train a DNN model without transferring the collected data into the cloud. We illustrate the applicability of the processing time of activity recognition on edge devices. We use a hierarchical model in which a global model is generated in the cloud, without requiring the raw data, and local models are trained on edge devices. The activity inference accuracy of the global model converges to a sufficient level after a few rounds of communication between edge devices and the cloud.
使用深度学习和传感器数据的活动识别可以帮助监测日常生活中需要帮助的人的活动和健康状况。深度神经网络(DNN)模型推断活动需要由家庭传感设备收集的数据。这些数据通常被发送到一个集中的云,用于训练模型。集中数据会带来隐私风险。收集的数据包含有关主题的敏感信息。基于云的方法增加了数据在没有所有者控制的情况下被存储和重用用于其他目的的风险。我们提出了一个系统,该系统使用边缘设备在本地实现活动和健康监测,并应用联邦学习来促进培训过程。这些设备使用Databox平台来管理在人们家中收集的传感器数据,在本地进行活动识别,并协同训练DNN模型,而无需将收集的数据传输到云端。我们说明了活动识别处理时间在边缘设备上的适用性。我们使用分层模型,在云中生成全局模型,不需要原始数据,而在边缘设备上训练局部模型。在边缘设备与云之间进行几轮通信后,全局模型的活动推理精度收敛到足够的水平。
{"title":"Privacy-preserving activity and health monitoring on databox","authors":"Yuchen Zhao, H. Haddadi, Severin Skillman, Shirin Enshaeifar, P. Barnaghi","doi":"10.1145/3378679.3394529","DOIUrl":"https://doi.org/10.1145/3378679.3394529","url":null,"abstract":"Activity recognition using deep learning and sensor data can help monitor activities and health conditions of people who need assistance in their daily lives. Deep Neural Network (DNN) models to infer the activities require data collected by in-home sensory devices. These data are often sent to a centralised cloud to be used for training the model. Centralising the data introduces privacy risks. The collected data contain sensitive information about the subjects. The cloud-based approach increases the risk that the data be stored and reused for other purposes without the owner's control. We propose a system that uses edge devices to implement activity and health monitoring locally and applies federated learning to facilitate the training process. The devices use the Databox platform to manage sensor data collected in people's homes, conduct activity recognition locally, and collaboratively train a DNN model without transferring the collected data into the cloud. We illustrate the applicability of the processing time of activity recognition on edge devices. We use a hierarchical model in which a global model is generated in the cloud, without requiring the raw data, and local models are trained on edge devices. The activity inference accuracy of the global model converges to a sufficient level after a few rounds of communication between edge devices and the cloud.","PeriodicalId":268360,"journal":{"name":"Proceedings of the Third ACM International Workshop on Edge Systems, Analytics and Networking","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128443253","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
期刊
Proceedings of the Third ACM International Workshop on Edge Systems, Analytics and Networking
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1