首页 > 最新文献

arXiv - CS - Emerging Technologies最新文献

英文 中文
Analysing Attacks on Blockchain Systems in a Layer-based Approach 用基于层的方法分析对区块链系统的攻击
Pub Date : 2024-09-16 DOI: arxiv-2409.10109
Joydip Das, Syed Ashraf Al Tasin, Md. Forhad Rabbi, Md Sadek Ferdous
Blockchain is a growing decentralized system built for transparency andimmutability. There have been several major attacks on blockchain-basedsystems, leaving a gap in the trustability of this system. This articlepresents a comprehensive study of 23 attacks on blockchain systems andcategorizes them using a layer-based approach. This approach provides anin-depth analysis of the feasibility and motivation of these attacks. Inaddition, a framework is proposed that enables a systematic analysis of theimpact and interconnection of these attacks, thereby providing a means ofidentifying potential attack vectors and designing appropriate countermeasuresto strengthen any blockchain system.
区块链是一个不断发展的去中心化系统,具有透明性和不变性。基于区块链的系统受到了几次重大攻击,使该系统的可信任度出现了漏洞。本文全面研究了针对区块链系统的 23 种攻击,并采用基于层的方法对其进行了分类。这种方法深入分析了这些攻击的可行性和动机。此外,本文还提出了一个框架,可以对这些攻击的影响和相互联系进行系统分析,从而为识别潜在的攻击载体和设计适当的应对措施提供一种手段,以加强任何区块链系统。
{"title":"Analysing Attacks on Blockchain Systems in a Layer-based Approach","authors":"Joydip Das, Syed Ashraf Al Tasin, Md. Forhad Rabbi, Md Sadek Ferdous","doi":"arxiv-2409.10109","DOIUrl":"https://doi.org/arxiv-2409.10109","url":null,"abstract":"Blockchain is a growing decentralized system built for transparency and\u0000immutability. There have been several major attacks on blockchain-based\u0000systems, leaving a gap in the trustability of this system. This article\u0000presents a comprehensive study of 23 attacks on blockchain systems and\u0000categorizes them using a layer-based approach. This approach provides an\u0000in-depth analysis of the feasibility and motivation of these attacks. In\u0000addition, a framework is proposed that enables a systematic analysis of the\u0000impact and interconnection of these attacks, thereby providing a means of\u0000identifying potential attack vectors and designing appropriate countermeasures\u0000to strengthen any blockchain system.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"19 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142263116","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pennsieve - A Collaborative Platform for Translational Neuroscience and Beyond Pennsieve--转化神经科学及其他领域的合作平台
Pub Date : 2024-09-16 DOI: arxiv-2409.10509
Zack GoldblumUniversity of Pennsylvania, Zhongchuan XuUniversity of Pennsylvania, Haoer ShiUniversity of Pennsylvania, Patryk OrzechowskiUniversity of PennsylvaniaAGH University of Krakow, Jamaal SpenceUniversity of Pennsylvania, Kathryn A DavisUniversity of Pennsylvania, Brian LittUniversity of Pennsylvania, Nishant SinhaUniversity of Pennsylvania, Joost WagenaarUniversity of Pennsylvania
The exponential growth of neuroscientific data necessitates platforms thatfacilitate data management and multidisciplinary collaboration. In this paper,we introduce Pennsieve - an open-source, cloud-based scientific data managementplatform built to meet these needs. Pennsieve supports complex multimodaldatasets and provides tools for data visualization and analyses. It takes acomprehensive approach to data integration, enabling researchers to definecustom metadata schemas and utilize advanced tools to filter and query theirdata. Pennsieve's modular architecture allows external applications to extendits capabilities, and collaborative workspaces with peer-reviewed datapublishing mechanisms promote high-quality datasets optimized for downstreamanalysis, both in the cloud and on-premises. Pennsieve forms the core for major neuroscience research programs includingthe NIH SPARC Initiative, NIH HEAL Initiative's PRECISION Human Pain Network,and NIH HEAL RE-JOIN Initiative. It serves more than 80 research groupsworldwide, along with several large-scale, inter-institutional projects atclinical sites through the University of Pennsylvania. Underpinning theSPARC.Science, Epilepsy.Science, and Pennsieve Discover portals, Pennsievestores over 125 TB of scientific data, with 35 TB of data publicly availableacross more than 350 high-impact datasets. It adheres to the findable,accessible, interoperable, and reusable (FAIR) principles of data sharing andis recognized as one of the NIH-approved Data Repositories. By facilitatingscientific data management, discovery, and analysis, Pennsieve fosters a robustand collaborative research ecosystem for neuroscience and beyond.
神经科学数据的指数级增长需要能促进数据管理和多学科协作的平台。在本文中,我们介绍了Pennsieve--一个为满足这些需求而构建的开源、基于云的科学数据管理平台。Pennsieve 支持复杂的多模式数据集,并提供数据可视化和分析工具。它采用综合方法进行数据整合,使研究人员能够定义自定义元数据模式,并利用高级工具过滤和查询他们的数据。Pennsieve的模块化架构允许外部应用程序扩展其功能,而具有同行评审数据发布机制的协作式工作空间则促进了云端和企业内部的高质量数据集,优化了下游分析。Pennsieve 是主要神经科学研究计划的核心,包括美国国立卫生研究院 SPARC 计划、美国国立卫生研究院 HEAL 计划的 PRECISION 人类疼痛网络和美国国立卫生研究院 HEAL RE-JOIN 计划。它为全球 80 多个研究小组提供服务,并通过宾夕法尼亚大学在临床基地开展了多个大型跨机构项目。宾夕法尼亚大学科学数据中心拥有超过 125 TB 的科学数据,其中 35 TB 的数据可通过 350 多个高影响力的数据集公开获取,这些数据支撑着SPARC.Science、Epilepsy.Science 和 Pennsievest Discover 门户网站。它遵循数据共享的可查找、可访问、可互操作和可重用(FAIR)原则,是美国国立卫生研究院(NIH)认可的数据存储库之一。通过促进科学数据的管理、发现和分析,Pennsieve 为神经科学及其他领域建立了一个强大的合作研究生态系统。
{"title":"Pennsieve - A Collaborative Platform for Translational Neuroscience and Beyond","authors":"Zack GoldblumUniversity of Pennsylvania, Zhongchuan XuUniversity of Pennsylvania, Haoer ShiUniversity of Pennsylvania, Patryk OrzechowskiUniversity of PennsylvaniaAGH University of Krakow, Jamaal SpenceUniversity of Pennsylvania, Kathryn A DavisUniversity of Pennsylvania, Brian LittUniversity of Pennsylvania, Nishant SinhaUniversity of Pennsylvania, Joost WagenaarUniversity of Pennsylvania","doi":"arxiv-2409.10509","DOIUrl":"https://doi.org/arxiv-2409.10509","url":null,"abstract":"The exponential growth of neuroscientific data necessitates platforms that\u0000facilitate data management and multidisciplinary collaboration. In this paper,\u0000we introduce Pennsieve - an open-source, cloud-based scientific data management\u0000platform built to meet these needs. Pennsieve supports complex multimodal\u0000datasets and provides tools for data visualization and analyses. It takes a\u0000comprehensive approach to data integration, enabling researchers to define\u0000custom metadata schemas and utilize advanced tools to filter and query their\u0000data. Pennsieve's modular architecture allows external applications to extend\u0000its capabilities, and collaborative workspaces with peer-reviewed data\u0000publishing mechanisms promote high-quality datasets optimized for downstream\u0000analysis, both in the cloud and on-premises. Pennsieve forms the core for major neuroscience research programs including\u0000the NIH SPARC Initiative, NIH HEAL Initiative's PRECISION Human Pain Network,\u0000and NIH HEAL RE-JOIN Initiative. It serves more than 80 research groups\u0000worldwide, along with several large-scale, inter-institutional projects at\u0000clinical sites through the University of Pennsylvania. Underpinning the\u0000SPARC.Science, Epilepsy.Science, and Pennsieve Discover portals, Pennsieve\u0000stores over 125 TB of scientific data, with 35 TB of data publicly available\u0000across more than 350 high-impact datasets. It adheres to the findable,\u0000accessible, interoperable, and reusable (FAIR) principles of data sharing and\u0000is recognized as one of the NIH-approved Data Repositories. By facilitating\u0000scientific data management, discovery, and analysis, Pennsieve fosters a robust\u0000and collaborative research ecosystem for neuroscience and beyond.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"18 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142263115","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Detection Made Easy: Potentials of Large Language Models for Solidity Vulnerabilities 轻松检测:大型语言模型在解决固体漏洞方面的潜力
Pub Date : 2024-09-15 DOI: arxiv-2409.10574
Md Tauseef Alam, Raju Halder, Abyayananda Maiti
The large-scale deployment of Solidity smart contracts on the Ethereummainnet has increasingly attracted financially-motivated attackers in recentyears. A few now-infamous attacks in Ethereum's history includes DAO attack in2016 (50 million dollars lost), Parity Wallet hack in 2017 (146 million dollarslocked), Beautychain's token BEC in 2018 (900 million dollars market value fellto 0), and NFT gaming blockchain breach in 2022 ($600 million in Ether stolen).This paper presents a comprehensive investigation of the use of large languagemodels (LLMs) and their capabilities in detecting OWASP Top Ten vulnerabilitiesin Solidity. We introduce a novel, class-balanced, structured, and labeleddataset named VulSmart, which we use to benchmark and compare the performanceof open-source LLMs such as CodeLlama, Llama2, CodeT5 and Falcon, alongsideclosed-source models like GPT-3.5 Turbo and GPT-4o Mini. Our proposed SmartVDframework is rigorously tested against these models through extensive automatedand manual evaluations, utilizing BLEU and ROUGE metrics to assess theeffectiveness of vulnerability detection in smart contracts. We also explorethree distinct prompting strategies-zero-shot, few-shot, andchain-of-thought-to evaluate the multi-class classification and generativecapabilities of the SmartVD framework. Our findings reveal that SmartVDoutperforms its open-source counterparts and even exceeds the performance ofclosed-source base models like GPT-3.5 and GPT-4 Mini. After fine-tuning, theclosed-source models, GPT-3.5 Turbo and GPT-4o Mini, achieved remarkableperformance with 99% accuracy in detecting vulnerabilities, 94% in identifyingtheir types, and 98% in determining severity. Notably, SmartVD performs bestwith the `chain-of-thought' prompting technique, whereas the fine-tunedclosed-source models excel with the `zero-shot' prompting approach.
近年来,以太坊主网上 Solidity 智能合约的大规模部署越来越多地吸引了出于经济动机的攻击者。以太坊历史上几起著名的攻击事件包括 2016 年的 DAO 攻击(损失 5000 万美元)、2017 年的 Parity Wallet 黑客攻击(1.46 亿美元被锁定)、2018 年的 Beautychain 代币 BEC(9 亿美元市值跌至 0)以及 2022 年的 NFT 游戏区块链漏洞(6 亿美元以太币被盗)。我们引入了一个名为 VulSmart 的新颖、类平衡、结构化和标签化数据集,并利用它对 CodeLlama、Llama2、CodeT5 和 Falcon 等开源 LLM,以及 GPT-3.5 Turbo 和 GPT-4o Mini 等封闭源模型的性能进行了基准测试和比较。我们提出的 SmartVD 框架通过广泛的自动和手动评估针对这些模型进行了严格测试,利用 BLEU 和 ROUGE 指标来评估智能合约中漏洞检测的有效性。我们还探索了三种不同的提示策略--零枪、少枪和思维链,以评估 SmartVD 框架的多类分类和生成能力。我们的研究结果表明,SmartVD 的性能优于其开源模型,甚至超过了 GPT-3.5 和 GPT-4 Mini 等闭源基础模型。经过微调后,GPT-3.5 Turbo 和 GPT-4o Mini 等闭源模型取得了显著的性能,检测漏洞的准确率达到 99%,识别漏洞类型的准确率达到 94%,判断漏洞严重性的准确率达到 98%。值得注意的是,SmartVD 在使用 "思维链 "提示技术时表现最佳,而经过微调的闭源模型在使用 "零镜头 "提示方法时表现出色。
{"title":"Detection Made Easy: Potentials of Large Language Models for Solidity Vulnerabilities","authors":"Md Tauseef Alam, Raju Halder, Abyayananda Maiti","doi":"arxiv-2409.10574","DOIUrl":"https://doi.org/arxiv-2409.10574","url":null,"abstract":"The large-scale deployment of Solidity smart contracts on the Ethereum\u0000mainnet has increasingly attracted financially-motivated attackers in recent\u0000years. A few now-infamous attacks in Ethereum's history includes DAO attack in\u00002016 (50 million dollars lost), Parity Wallet hack in 2017 (146 million dollars\u0000locked), Beautychain's token BEC in 2018 (900 million dollars market value fell\u0000to 0), and NFT gaming blockchain breach in 2022 ($600 million in Ether stolen).\u0000This paper presents a comprehensive investigation of the use of large language\u0000models (LLMs) and their capabilities in detecting OWASP Top Ten vulnerabilities\u0000in Solidity. We introduce a novel, class-balanced, structured, and labeled\u0000dataset named VulSmart, which we use to benchmark and compare the performance\u0000of open-source LLMs such as CodeLlama, Llama2, CodeT5 and Falcon, alongside\u0000closed-source models like GPT-3.5 Turbo and GPT-4o Mini. Our proposed SmartVD\u0000framework is rigorously tested against these models through extensive automated\u0000and manual evaluations, utilizing BLEU and ROUGE metrics to assess the\u0000effectiveness of vulnerability detection in smart contracts. We also explore\u0000three distinct prompting strategies-zero-shot, few-shot, and\u0000chain-of-thought-to evaluate the multi-class classification and generative\u0000capabilities of the SmartVD framework. Our findings reveal that SmartVD\u0000outperforms its open-source counterparts and even exceeds the performance of\u0000closed-source base models like GPT-3.5 and GPT-4 Mini. After fine-tuning, the\u0000closed-source models, GPT-3.5 Turbo and GPT-4o Mini, achieved remarkable\u0000performance with 99% accuracy in detecting vulnerabilities, 94% in identifying\u0000their types, and 98% in determining severity. Notably, SmartVD performs best\u0000with the `chain-of-thought' prompting technique, whereas the fine-tuned\u0000closed-source models excel with the `zero-shot' prompting approach.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"41 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142269893","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Exploring Utility in a Real-World Warehouse Optimization Problem: Formulation Based on Quantun Annealers and Preliminary Results 在现实世界的仓库优化问题中探索实用性:基于 Quantun 退火炉的表述和初步结果
Pub Date : 2024-09-15 DOI: arxiv-2409.09706
Eneko Osaba, Esther Villar-Rodriguez, Antón Asla
In the current NISQ-era, one of the major challenges faced by researchers andpractitioners lies in figuring out how to combine quantum and classicalcomputing in the most efficient and innovative way. In this paper, we present amechanism coined as Quantum Initialization for Warehouse Optimization Problemthat resorts to D-Wave's Quantum Annealer. The module has been specificallydesigned to be embedded into already existing classical software dedicated tothe optimization of a real-world industrial problem. We preliminary tested theimplemented mechanism through a two-phase experiment against the classicalversion of the software.
在当前的 NISQ 时代,研究人员和实践者面临的主要挑战之一是如何以最高效、最创新的方式将量子计算与经典计算结合起来。在本文中,我们提出了一种被称为仓库优化问题量子初始化(Quantum Initialization for Warehouse Optimization Problem)的机制,它采用了 D-Wave 的量子退火器(Quantum Annealer)。该模块经过专门设计,可以嵌入到现有的经典软件中,专门用于优化现实世界中的工业问题。我们通过与经典版本软件的两阶段实验,对实施的机制进行了初步测试。
{"title":"Exploring Utility in a Real-World Warehouse Optimization Problem: Formulation Based on Quantun Annealers and Preliminary Results","authors":"Eneko Osaba, Esther Villar-Rodriguez, Antón Asla","doi":"arxiv-2409.09706","DOIUrl":"https://doi.org/arxiv-2409.09706","url":null,"abstract":"In the current NISQ-era, one of the major challenges faced by researchers and\u0000practitioners lies in figuring out how to combine quantum and classical\u0000computing in the most efficient and innovative way. In this paper, we present a\u0000mechanism coined as Quantum Initialization for Warehouse Optimization Problem\u0000that resorts to D-Wave's Quantum Annealer. The module has been specifically\u0000designed to be embedded into already existing classical software dedicated to\u0000the optimization of a real-world industrial problem. We preliminary tested the\u0000implemented mechanism through a two-phase experiment against the classical\u0000version of the software.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"23 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142263113","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
High Definition Map Mapping and Update: A General Overview and Future Directions 高清地图制图与更新:概览与未来方向
Pub Date : 2024-09-15 DOI: arxiv-2409.09726
Benny Wijaya, Kun Jiang, Mengmeng Yang, Tuopu Wen, Yunlong Wang, Xuewei Tang, Zheng Fu, Taohua Zhou, Diange Yang
Along with the rapid growth of autonomous vehicles (AVs), more and moredemands are required for environment perception technology. Among others, HDmapping has become one of the more prominent roles in helping the vehiclerealize essential tasks such as localization and path planning. Whileincreasing research efforts have been directed toward HD Map development.However, a comprehensive overview of the overall HD map mapping and updateframework is still lacking. This article introduces the development and currentstate of the algorithm involved in creating HD map mapping and its maintenance.As part of this study, the primary data preprocessing approach of processingraw data to information ready to feed for mapping and update purposes, semanticsegmentation, and localization are also briefly reviewed. Moreover, the maptaxonomy, ontology, and quality assessment are extensively discussed, the mapdata's general representation method is presented, and the mapping algorithmranging from SLAM to transformers learning-based approaches are also discussed.The development of the HD map update algorithm, from change detection to theupdate methods, is also presented. Finally, the authors discuss possible futuredevelopments and the remaining challenges in HD map mapping and updatetechnology. This paper simultaneously serves as a position paper and tutorialto those new to HD map mapping and update domains.
随着自动驾驶汽车(AV)的快速发展,对环境感知技术的要求也越来越高。其中,高清地图在帮助车辆实现定位和路径规划等基本任务方面的作用尤为突出。虽然越来越多的研究人员致力于高清地图的开发,但目前仍缺乏对整个高清地图绘制和更新框架工作的全面概述。本文介绍了创建高清地图制图及其维护所涉及的算法的发展和现状。作为这项研究的一部分,本文还简要回顾了将raw数据处理为可用于制图和更新目的的信息、语义分割和定位的主要数据预处理方法。此外,还广泛讨论了地图分类学、本体论和质量评估,介绍了地图数据的一般表示方法,并讨论了从 SLAM 到基于转换器学习方法的绘图算法。最后,作者讨论了高清地图制图和更新技术未来可能的发展和仍然面临的挑战。本文既是一篇立场论文,也是对高清地图制图和更新领域新手的指导。
{"title":"High Definition Map Mapping and Update: A General Overview and Future Directions","authors":"Benny Wijaya, Kun Jiang, Mengmeng Yang, Tuopu Wen, Yunlong Wang, Xuewei Tang, Zheng Fu, Taohua Zhou, Diange Yang","doi":"arxiv-2409.09726","DOIUrl":"https://doi.org/arxiv-2409.09726","url":null,"abstract":"Along with the rapid growth of autonomous vehicles (AVs), more and more\u0000demands are required for environment perception technology. Among others, HD\u0000mapping has become one of the more prominent roles in helping the vehicle\u0000realize essential tasks such as localization and path planning. While\u0000increasing research efforts have been directed toward HD Map development.\u0000However, a comprehensive overview of the overall HD map mapping and update\u0000framework is still lacking. This article introduces the development and current\u0000state of the algorithm involved in creating HD map mapping and its maintenance.\u0000As part of this study, the primary data preprocessing approach of processing\u0000raw data to information ready to feed for mapping and update purposes, semantic\u0000segmentation, and localization are also briefly reviewed. Moreover, the map\u0000taxonomy, ontology, and quality assessment are extensively discussed, the map\u0000data's general representation method is presented, and the mapping algorithm\u0000ranging from SLAM to transformers learning-based approaches are also discussed.\u0000The development of the HD map update algorithm, from change detection to the\u0000update methods, is also presented. Finally, the authors discuss possible future\u0000developments and the remaining challenges in HD map mapping and update\u0000technology. This paper simultaneously serves as a position paper and tutorial\u0000to those new to HD map mapping and update domains.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"4 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142263117","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Quantum data encoding as a distinct abstraction layer in the design of quantum circuits 量子数据编码作为量子电路设计中的一个独特抽象层
Pub Date : 2024-09-14 DOI: arxiv-2409.09339
Gabriele Agliardi, Enrico Prati
Complex quantum circuits are constituted by combinations of quantumsubroutines. The computation is possible as long as the quantum data encodingis consistent throughout the circuit. Despite its fundamental importance, theformalization of quantum data encoding has never been addressed systematicallyso far. We formalize the concept of quantum data encoding, namely the formatproviding a representation of a data set through a quantum state, as a distinctabstract layer with respect to the associated data loading circuit. We surveyexisting encoding methods and their respective strategies forclassical-to-quantum exact and approximate data loading, for thequantum-to-classical extraction of information from states, and forquantum-to-quantum encoding conversion. Next, we show how major quantumalgorithms find a natural interpretation in terms of data loading. Forinstance, the Quantum Fourier Transform is described as a quantum encodingconverter, while the Quantum Amplitude Estimation as an extraction routine. Thenew conceptual framework is exemplified by considering its application toquantum-based Monte Carlo simulations, thus showcasing the power of theproposed formalism for the description of complex quantum circuits. Indeed, theapproach clarifies the structure of complex quantum circuits and enables theirefficient design.
复杂的量子电路由量子子程序组合而成。只要量子数据编码在整个电路中保持一致,计算就有可能实现。尽管量子数据编码的形式化具有根本性的重要意义,但迄今为止从未有人系统地解决过这一问题。我们将量子数据编码的概念形式化,即通过量子态提供数据集表示的格式,作为相关数据加载电路的一个独特抽象层。我们研究了现有的编码方法及其各自的策略,包括从经典到量子的精确和近似数据加载、从量子到经典的状态信息提取,以及从量子到量子的编码转换。接下来,我们将展示主要量子算法如何在数据加载方面找到自然解释。例如,量子傅里叶变换被描述为一种量子编码转换器,而量子振幅估计则是一种提取程序。通过将新概念框架应用于基于量子的蒙特卡罗模拟,展示了所提出的形式主义在描述复杂量子电路方面的威力。事实上,该方法阐明了复杂量子电路的结构,并实现了高效设计。
{"title":"Quantum data encoding as a distinct abstraction layer in the design of quantum circuits","authors":"Gabriele Agliardi, Enrico Prati","doi":"arxiv-2409.09339","DOIUrl":"https://doi.org/arxiv-2409.09339","url":null,"abstract":"Complex quantum circuits are constituted by combinations of quantum\u0000subroutines. The computation is possible as long as the quantum data encoding\u0000is consistent throughout the circuit. Despite its fundamental importance, the\u0000formalization of quantum data encoding has never been addressed systematically\u0000so far. We formalize the concept of quantum data encoding, namely the format\u0000providing a representation of a data set through a quantum state, as a distinct\u0000abstract layer with respect to the associated data loading circuit. We survey\u0000existing encoding methods and their respective strategies for\u0000classical-to-quantum exact and approximate data loading, for the\u0000quantum-to-classical extraction of information from states, and for\u0000quantum-to-quantum encoding conversion. Next, we show how major quantum\u0000algorithms find a natural interpretation in terms of data loading. For\u0000instance, the Quantum Fourier Transform is described as a quantum encoding\u0000converter, while the Quantum Amplitude Estimation as an extraction routine. The\u0000new conceptual framework is exemplified by considering its application to\u0000quantum-based Monte Carlo simulations, thus showcasing the power of the\u0000proposed formalism for the description of complex quantum circuits. Indeed, the\u0000approach clarifies the structure of complex quantum circuits and enables their\u0000efficient design.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"27 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142263114","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ProcessTBench: An LLM Plan Generation Dataset for Process Mining ProcessTBench:用于流程挖掘的 LLM 计划生成数据集
Pub Date : 2024-09-13 DOI: arxiv-2409.09191
Andrei Cosmin Redis, Mohammadreza Fani Sani, Bahram Zarrin, Andrea Burattin
Large Language Models (LLMs) have shown significant promise in plangeneration. Yet, existing datasets often lack the complexity needed foradvanced tool use scenarios - such as handling paraphrased query statements,supporting multiple languages, and managing actions that can be done inparallel. These scenarios are crucial for evaluating the evolving capabilitiesof LLMs in real-world applications. Moreover, current datasets don't enable thestudy of LLMs from a process perspective, particularly in scenarios whereunderstanding typical behaviors and challenges in executing the same processunder different conditions or formulations is crucial. To address these gaps,we present the ProcessTBench dataset, an extension of the TaskBench datasetspecifically designed to evaluate LLMs within a process mining framework.
大型语言模型(LLMs)在 plangeneration 方面显示出了巨大的前景。然而,现有的数据集往往缺乏高级工具使用场景所需的复杂性,例如处理解析查询语句、支持多种语言以及管理可并行执行的操作。这些场景对于评估 LLM 在实际应用中不断发展的能力至关重要。此外,当前的数据集无法从流程的角度来研究 LLM,尤其是在一些场景中,了解在不同条件或配方下执行相同流程的典型行为和挑战至关重要。为了弥补这些不足,我们提出了 ProcessTBench 数据集,它是 TaskBench 数据集的扩展,专门用于在流程挖掘框架内评估 LLM。
{"title":"ProcessTBench: An LLM Plan Generation Dataset for Process Mining","authors":"Andrei Cosmin Redis, Mohammadreza Fani Sani, Bahram Zarrin, Andrea Burattin","doi":"arxiv-2409.09191","DOIUrl":"https://doi.org/arxiv-2409.09191","url":null,"abstract":"Large Language Models (LLMs) have shown significant promise in plan\u0000generation. Yet, existing datasets often lack the complexity needed for\u0000advanced tool use scenarios - such as handling paraphrased query statements,\u0000supporting multiple languages, and managing actions that can be done in\u0000parallel. These scenarios are crucial for evaluating the evolving capabilities\u0000of LLMs in real-world applications. Moreover, current datasets don't enable the\u0000study of LLMs from a process perspective, particularly in scenarios where\u0000understanding typical behaviors and challenges in executing the same process\u0000under different conditions or formulations is crucial. To address these gaps,\u0000we present the ProcessTBench dataset, an extension of the TaskBench dataset\u0000specifically designed to evaluate LLMs within a process mining framework.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"45 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142263118","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Delay Balancing with Clock-Follow-Data: Optimizing Area Delay Trade-offs for Robust Rapid Single Flux Quantum Circuits 利用时钟跟随数据实现延迟平衡:优化稳健快速单通量量子电路的面积延迟权衡
Pub Date : 2024-09-08 DOI: arxiv-2409.04944
Robert S. Aviles, Phalgun G K, Peter A. Beerel
This paper proposes an algorithm for synthesis of clock-follow-data designsthat provides robustness against timing violations for RSFQ circuits whilemaintaining high performance and minimizing area costs. Since superconductinglogic gates must be clocked, managing data flow is a challenging problem thatoften requires the insertion of many path balancing D Flips (DFFs) to properlysequence data, leading to a substantial increase in area. To address thischallenge, we present an algorithm to insert DFFs into clock-follow-data RSFQcircuits that partially balances the delays within the circuit to achieve atarget throughput while minimizing area. Our algorithm can account for expectedtiming variations and, by adjusting the bias of the clock network and clockfrequency, we can mitigate unexpected timing violations post-fabrication.Quantifying the benefits of our approach with a benchmark suite with nominaldelays, our designs offer an average 1.48x improvement in area delay product(ADP) over high frequency full path balancing (FPB) designs and a 2.07ximprovement in ADP over the state of the art robust circuits provided bystate-of-the-art (SOTA) multi-phase clocking solutions.
本文提出了一种用于合成时钟跟随数据设计的算法,该算法在保持高性能和最小面积成本的同时,还能防止 RSFQ 电路出现时序违规。由于超导逻辑门必须时钟化,因此管理数据流是一个具有挑战性的问题,通常需要插入许多路径平衡 D Flips (DFF),以正确排序数据,从而导致面积大幅增加。为解决这一难题,我们提出了一种在时钟跟随数据 RSFQ 电路中插入 DFF 的算法,该算法可部分平衡电路内的延迟,从而在实现目标吞吐量的同时最大限度地减少面积。我们的算法可以考虑预期的时序变化,通过调整时钟网络的偏置和时钟频率,我们可以在制造后减轻意外的时序违规。我们的设计与高频全路径平衡(FPB)设计相比,在面积延迟积(ADP)方面平均提高了 1.48 倍,与最先进的多相时钟解决方案(SOTA)所提供的最先进的稳健电路相比,在面积延迟积(ADP)方面提高了 2.07 倍。
{"title":"Delay Balancing with Clock-Follow-Data: Optimizing Area Delay Trade-offs for Robust Rapid Single Flux Quantum Circuits","authors":"Robert S. Aviles, Phalgun G K, Peter A. Beerel","doi":"arxiv-2409.04944","DOIUrl":"https://doi.org/arxiv-2409.04944","url":null,"abstract":"This paper proposes an algorithm for synthesis of clock-follow-data designs\u0000that provides robustness against timing violations for RSFQ circuits while\u0000maintaining high performance and minimizing area costs. Since superconducting\u0000logic gates must be clocked, managing data flow is a challenging problem that\u0000often requires the insertion of many path balancing D Flips (DFFs) to properly\u0000sequence data, leading to a substantial increase in area. To address this\u0000challenge, we present an algorithm to insert DFFs into clock-follow-data RSFQ\u0000circuits that partially balances the delays within the circuit to achieve a\u0000target throughput while minimizing area. Our algorithm can account for expected\u0000timing variations and, by adjusting the bias of the clock network and clock\u0000frequency, we can mitigate unexpected timing violations post-fabrication.\u0000Quantifying the benefits of our approach with a benchmark suite with nominal\u0000delays, our designs offer an average 1.48x improvement in area delay product\u0000(ADP) over high frequency full path balancing (FPB) designs and a 2.07x\u0000improvement in ADP over the state of the art robust circuits provided by\u0000state-of-the-art (SOTA) multi-phase clocking solutions.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"282 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142214100","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
3D System Design: A Case for Building Customized Modular Systems in 3D 三维系统设计:以三维方式构建定制模块化系统的案例
Pub Date : 2024-09-06 DOI: arxiv-2409.09068
Philip Emma, Eren Kurshan
3D promises a new dimension in composing systems by aggregating chips.Literally. While the most common uses are still tightly connected with itsearly forms as a packaging technology, new application domains have beenemerging. As the underlying technology continues to evolve, the uniqueleverages of 3D have become increasingly appealing to a larger range ofapplications: from embedded mobile applications to servers and memory systems.In this paper we focus on the system-level implications of 3D technology,trying to differentiate the unique advantages that it provides to differentmarket segments and applications.
三维技术通过将芯片聚合在一起,为系统构成带来了新的前景。尽管三维技术最常见的用途仍与其早期的封装技术密切相关,但新的应用领域正在不断涌现。随着底层技术的不断发展,3D 技术的独特优势对更广泛的应用领域越来越有吸引力:从嵌入式移动应用到服务器和内存系统。在本文中,我们将重点关注 3D 技术在系统层面的影响,试图区分 3D 技术为不同细分市场和应用提供的独特优势。
{"title":"3D System Design: A Case for Building Customized Modular Systems in 3D","authors":"Philip Emma, Eren Kurshan","doi":"arxiv-2409.09068","DOIUrl":"https://doi.org/arxiv-2409.09068","url":null,"abstract":"3D promises a new dimension in composing systems by aggregating chips.\u0000Literally. While the most common uses are still tightly connected with its\u0000early forms as a packaging technology, new application domains have been\u0000emerging. As the underlying technology continues to evolve, the unique\u0000leverages of 3D have become increasingly appealing to a larger range of\u0000applications: from embedded mobile applications to servers and memory systems.\u0000In this paper we focus on the system-level implications of 3D technology,\u0000trying to differentiate the unique advantages that it provides to different\u0000market segments and applications.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"48 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142269894","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LLM-based event abstraction and integration for IoT-sourced logs 基于 LLM 的事件抽象和整合物联网日志
Pub Date : 2024-09-05 DOI: arxiv-2409.03478
Mohsen Shirali, Mohammadreza Fani Sani, Zahra Ahmadi, Estefania Serral
The continuous flow of data collected by Internet of Things (IoT) devices,has revolutionised our ability to understand and interact with the world acrossvarious applications. However, this data must be prepared and transformed intoevent data before analysis can begin. In this paper, we shed light on thepotential of leveraging Large Language Models (LLMs) in event abstraction andintegration. Our approach aims to create event records from raw sensor readingsand merge the logs from multiple IoT sources into a single event log suitablefor further Process Mining applications. We demonstrate the capabilities ofLLMs in event abstraction considering a case study for IoT application inelderly care and longitudinal health monitoring. The results, showing onaverage an accuracy of 90% in detecting high-level activities. These resultshighlight LLMs' promising potential in addressing event abstraction andintegration challenges, effectively bridging the existing gap.
物联网(IoT)设备收集的数据流源源不断,彻底改变了我们在各种应用中了解世界并与之互动的能力。然而,在开始分析之前,必须将这些数据准备好并转换成事件数据。在本文中,我们将阐明利用大型语言模型(LLM)进行事件抽象和整合的潜力。我们的方法旨在从原始传感器读数中创建事件记录,并将来自多个物联网源的日志合并为适合进一步流程挖掘应用的单一事件日志。我们通过一个物联网应用于老年人护理和纵向健康监测的案例研究,展示了LLMs 在事件抽象方面的能力。结果表明,检测高级活动的平均准确率为 90%。这些结果凸显了 LLMs 在解决事件抽象和集成挑战方面的巨大潜力,有效弥补了现有差距。
{"title":"LLM-based event abstraction and integration for IoT-sourced logs","authors":"Mohsen Shirali, Mohammadreza Fani Sani, Zahra Ahmadi, Estefania Serral","doi":"arxiv-2409.03478","DOIUrl":"https://doi.org/arxiv-2409.03478","url":null,"abstract":"The continuous flow of data collected by Internet of Things (IoT) devices,\u0000has revolutionised our ability to understand and interact with the world across\u0000various applications. However, this data must be prepared and transformed into\u0000event data before analysis can begin. In this paper, we shed light on the\u0000potential of leveraging Large Language Models (LLMs) in event abstraction and\u0000integration. Our approach aims to create event records from raw sensor readings\u0000and merge the logs from multiple IoT sources into a single event log suitable\u0000for further Process Mining applications. We demonstrate the capabilities of\u0000LLMs in event abstraction considering a case study for IoT application in\u0000elderly care and longitudinal health monitoring. The results, showing on\u0000average an accuracy of 90% in detecting high-level activities. These results\u0000highlight LLMs' promising potential in addressing event abstraction and\u0000integration challenges, effectively bridging the existing gap.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"40 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142214102","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
arXiv - CS - Emerging Technologies
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1