首页 > 最新文献

2021 IEEE Hot Chips 33 Symposium (HCS)最新文献

英文 中文
Nvidia Data Center Processing Unit (DPU) Architecture Nvidia数据中心处理单元(DPU)架构
Pub Date : 2021-08-22 DOI: 10.1109/HCS52781.2021.9567066
Idan Burstein
NVIDIA DPU Enables the Data Center as the New Unit of Computing The CPU can no longer do it all Must offload & isolate server infrastructure tasks to a DPU Effective DPU must offer hardware acceleration and security isolation To enable such effective DPU, need to develop broad software eco-system to utilize hardware acceleration across variety of disciplines (e.g. HPC, AI/ML, Storage, Networking, Security) - DOCA NVIDIA DPU & DOCA is a computing platform with rich stack optimized ideal for AI, bare metal cloud, cloud supercomputing, storage, gaming, 5G wireless, and more NVIDIA is committed to line rate performance every generation.
NVIDIA DPU使得数据中心的新单元计算CPU再也不能做一切必须卸载和隔离服务器基础设施任务DPU有效DPU必须提供硬件加速和安全隔离等使DPU有效,需要开发广泛的软件生态系统利用硬件加速在各种各样的学科(例如HPC, AI /毫升,存储、网络、安全)——DOCA NVIDIA DPU & DOCA堆栈是一个计算平台,丰富优化适合人工智能,裸机云、云超级计算、存储、游戏、5G无线等等,英伟达每一代都致力于线速性能。
{"title":"Nvidia Data Center Processing Unit (DPU) Architecture","authors":"Idan Burstein","doi":"10.1109/HCS52781.2021.9567066","DOIUrl":"https://doi.org/10.1109/HCS52781.2021.9567066","url":null,"abstract":"NVIDIA DPU Enables the Data Center as the New Unit of Computing The CPU can no longer do it all Must offload & isolate server infrastructure tasks to a DPU Effective DPU must offer hardware acceleration and security isolation To enable such effective DPU, need to develop broad software eco-system to utilize hardware acceleration across variety of disciplines (e.g. HPC, AI/ML, Storage, Networking, Security) - DOCA NVIDIA DPU & DOCA is a computing platform with rich stack optimized ideal for AI, bare metal cloud, cloud supercomputing, storage, gaming, 5G wireless, and more NVIDIA is committed to line rate performance every generation.","PeriodicalId":246531,"journal":{"name":"2021 IEEE Hot Chips 33 Symposium (HCS)","volume":"78 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126586932","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 34
Industry's First 7.2 Gbps 512GB DDR5 Module 业界首款7.2 Gbps 512GB DDR5模块
Pub Date : 2021-08-22 DOI: 10.1109/HCS52781.2021.9567190
Sung Joo Park, Jonghoon J. Kim, Kun Joo, Young-Ho Lee, Kyoungsun Kim, Young-Tae Kim, W. Na, I. Choi, Hye-Seung Yu, W. Kim, J. Jung, Jaejun Lee, Dohyung Kim, Young-Uk Chang, G. Han, Hangi-Jung, Sunwon Kang, Jeonghyeon Cho, H. Song, T. Oh, Y. Sohn, Sang-Wook Hwang, Jooyoung Lee
Spurred by the increasing market needs for big data and cloud services, global server suppliers and hyper- scalers are looking to adopt high-speed and large-capacity memory modules. To fulfill this trend, the brand- new low-voltage operable DDR5 (double data rate 5th generation) memory can be an appropriate solution, with the highest speed of 7.2 Gbps and the largest capacity of 512 GB. However, some critical obstacles, such as increased capacity and high-speed I/O requirements, unstable power noise occurrences, high power consumption, and increase in operating temperature, must be overcome. This poster will cover various technical pathfinding solutions for world's first DDR5 512 GB module with an advanced DRAM process and I/O schemes, package technology, and module architecture regarding improvements in the following four aspects: performance, speed, capacity, and power. This will unveil the industry's first high-performance and large-capacity memory product with 8-stacked DDR5 DRAMs. Samsung believes that this product will pave the way for achieving both higher bandwidth and lower power consumption to inaugurate the era of terabyte DRAM modules for next-gen servers.
受市场对大数据和云服务日益增长的需求的刺激,全球服务器供应商和超大规模厂商正在寻求采用高速和大容量的内存模块。为了满足这一趋势,全新的低电压可操作DDR5(双数据速率第5代)存储器可能是一个合适的解决方案,其最高速度为7.2 Gbps,最大容量为512 GB。但是,必须克服一些关键障碍,例如容量增加和高速I/O要求,不稳定的功率噪声,高功耗和工作温度升高。这张海报将介绍全球首款DDR5 512 GB模块的各种技术寻路方案,该模块采用先进的DRAM工艺和I/O方案,封装技术和模块架构,从性能、速度、容量和功耗四个方面进行改进。这将是业界首款采用8层DDR5 dram的高性能大容量内存产品。三星电子认为,该产品将为实现更高的带宽和更低的功耗铺平道路,从而开创新一代服务器的tb DRAM时代。
{"title":"Industry's First 7.2 Gbps 512GB DDR5 Module","authors":"Sung Joo Park, Jonghoon J. Kim, Kun Joo, Young-Ho Lee, Kyoungsun Kim, Young-Tae Kim, W. Na, I. Choi, Hye-Seung Yu, W. Kim, J. Jung, Jaejun Lee, Dohyung Kim, Young-Uk Chang, G. Han, Hangi-Jung, Sunwon Kang, Jeonghyeon Cho, H. Song, T. Oh, Y. Sohn, Sang-Wook Hwang, Jooyoung Lee","doi":"10.1109/HCS52781.2021.9567190","DOIUrl":"https://doi.org/10.1109/HCS52781.2021.9567190","url":null,"abstract":"Spurred by the increasing market needs for big data and cloud services, global server suppliers and hyper- scalers are looking to adopt high-speed and large-capacity memory modules. To fulfill this trend, the brand- new low-voltage operable DDR5 (double data rate 5th generation) memory can be an appropriate solution, with the highest speed of 7.2 Gbps and the largest capacity of 512 GB. However, some critical obstacles, such as increased capacity and high-speed I/O requirements, unstable power noise occurrences, high power consumption, and increase in operating temperature, must be overcome. This poster will cover various technical pathfinding solutions for world's first DDR5 512 GB module with an advanced DRAM process and I/O schemes, package technology, and module architecture regarding improvements in the following four aspects: performance, speed, capacity, and power. This will unveil the industry's first high-performance and large-capacity memory product with 8-stacked DDR5 DRAMs. Samsung believes that this product will pave the way for achieving both higher bandwidth and lower power consumption to inaugurate the era of terabyte DRAM modules for next-gen servers.","PeriodicalId":246531,"journal":{"name":"2021 IEEE Hot Chips 33 Symposium (HCS)","volume":"129 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132272938","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SOT-MRAM – Third generation MRAM memory opens new opportunities : Hot Chips Conference August 2021 SOT-MRAM -第三代MRAM存储器开启了新的机遇:热芯片会议2021年8月
Pub Date : 2021-08-22 DOI: 10.1109/HCS52781.2021.9567072
Barry A. Hoberman, J. Nozieres
SOT is a straightforward extension of today’s ‘in production’ MRAM technologies running in major foundries. First memory technology to genuinely have the capability to converge both SRAM and NVM characteristics in advanced CMOS nodes $( le 28$ nm). Power, cost, and performance benefits of SOT are compelling. SOT is still in development, but look for market visibility to begin around 2024
SOT是当今在主要铸造厂运行的“生产中”MRAM技术的直接扩展。第一个真正有能力在先进的CMOS节点上融合SRAM和NVM特性的存储技术( $ 28$ nm)。SOT的功率、成本和性能优势是引人注目的。SOT仍在开发中,但市场知名度将在2024年左右开始
{"title":"SOT-MRAM – Third generation MRAM memory opens new opportunities : Hot Chips Conference August 2021","authors":"Barry A. Hoberman, J. Nozieres","doi":"10.1109/HCS52781.2021.9567072","DOIUrl":"https://doi.org/10.1109/HCS52781.2021.9567072","url":null,"abstract":"SOT is a straightforward extension of today’s ‘in production’ MRAM technologies running in major foundries. First memory technology to genuinely have the capability to converge both SRAM and NVM characteristics in advanced CMOS nodes $( le 28$ nm). Power, cost, and performance benefits of SOT are compelling. SOT is still in development, but look for market visibility to begin around 2024","PeriodicalId":246531,"journal":{"name":"2021 IEEE Hot Chips 33 Symposium (HCS)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133619086","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
RDNA™ 2 Gaming Architecture RDNA™2游戏架构
Pub Date : 2021-08-22 DOI: 10.1109/HCS52781.2021.9567555
Andrew Pomianowski
This presentation contains forward-looking statements concerning Advanced Micro Devices, Inc. (AMD) including, but not limited to, the features, functionality, availability, timing, expectations and expected benefits of AMD future products, including Ryzen™ 5000 Series CPUs and Socket AM4, which are made pursuant to the Safe Harbor provisions of the Private Securities Litigation Reform Act of 1995. Forward-looking statements are commonly identified by words such as “would,“ “may,“ “expects,“ “believes,“ “plans,“ “intends,“ “projects“ and other terms with similar meaning. Investors are cautioned that the forward-looking statements in this presentation are based on current beliefs, assumptions and expectations, speak only as of the date of this presentation and involve risks and uncertainties that could cause actual results to differ materially from current expectations. Such statements are subject to certain known and unknown risks and uncertainties, many of which are difficult to predict and generally beyond AMD’s control, that could cause actual results and other future events to differ materially from those expressed in, or implied or projected by, the forward-looking information and statements. Investors are urged to review in detail the risks and uncertainties in AMD’s Securities and Exchange Commission filings, including but not limited to AMD’s Quarterly Report on Form 10-Qfrom the quarter ending on June 26, 2021.
本演示包含有关AMD的前瞻性陈述,包括但不限于AMD未来产品(包括Ryzen™5000系列cpu和Socket AM4)的特性、功能、可用性、时间、预期和预期优势,这些陈述是根据1995年《私人证券诉讼改革法案》的安全港条款制定的。前瞻性陈述通常用“会”、“可能”、“预期”、“相信”、“计划”、“打算”、“项目”和其他具有类似含义的术语来标识。投资者需注意,本报告中的前瞻性陈述基于当前的信念、假设和预期,仅在本报告发布之日发表,并涉及可能导致实际结果与当前预期存在重大差异的风险和不确定性。此类陈述受到某些已知和未知风险和不确定性的影响,其中许多风险和不确定性难以预测,并且通常超出AMD的控制范围,可能导致实际结果和其他未来事件与前瞻性信息和陈述中所表达、暗示或预测的内容存在重大差异。我们敦促投资者详细审查AMD提交给美国证券交易委员会的文件中的风险和不确定性,包括但不限于AMD截至2021年6月26日的季度10- qform季度报告。
{"title":"RDNA™ 2 Gaming Architecture","authors":"Andrew Pomianowski","doi":"10.1109/HCS52781.2021.9567555","DOIUrl":"https://doi.org/10.1109/HCS52781.2021.9567555","url":null,"abstract":"This presentation contains forward-looking statements concerning Advanced Micro Devices, Inc. (AMD) including, but not limited to, the features, functionality, availability, timing, expectations and expected benefits of AMD future products, including Ryzen™ 5000 Series CPUs and Socket AM4, which are made pursuant to the Safe Harbor provisions of the Private Securities Litigation Reform Act of 1995. Forward-looking statements are commonly identified by words such as “would,“ “may,“ “expects,“ “believes,“ “plans,“ “intends,“ “projects“ and other terms with similar meaning. Investors are cautioned that the forward-looking statements in this presentation are based on current beliefs, assumptions and expectations, speak only as of the date of this presentation and involve risks and uncertainties that could cause actual results to differ materially from current expectations. Such statements are subject to certain known and unknown risks and uncertainties, many of which are difficult to predict and generally beyond AMD’s control, that could cause actual results and other future events to differ materially from those expressed in, or implied or projected by, the forward-looking information and statements. Investors are urged to review in detail the risks and uncertainties in AMD’s Securities and Exchange Commission filings, including but not limited to AMD’s Quarterly Report on Form 10-Qfrom the quarter ending on June 26, 2021.","PeriodicalId":246531,"journal":{"name":"2021 IEEE Hot Chips 33 Symposium (HCS)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116780849","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Accelerating ML Recommendation with over a Thousand RISC-V/Tensor Processors on Esperanto’s ET-SoC-1 Chip Esperanto ET-SoC-1芯片上超过一千个RISC-V/张量处理器加速机器学习推荐
Pub Date : 2021-08-22 DOI: 10.1109/HCS52781.2021.9566904
D. Ditzel, R. Espasa, Nivard Aymerich, Allen Baum, Tom Berg, Jim Burr, Eric Hao, Jayesh Iyer, Miquel Izquierdo, Shankar Jayaratnam, Darren Jones, Chris Klingner, Jin Kim, Stephen Lee, Marc Lupon, G. Magklis, Bojan Maric, Rajib Nath, Michael Neilly, Duane J. Northcutt, Bill Orner, Jose Renau, Gerard Reves, X. Revés, Tom Riordan, Pedro Sanchez, S. Samudrala, Guillem Sole, Raymond Tang, Tommy Thorn, Francisco Torres, S. Tortella, Daniel Yau
The ET-SoC-1 has over a thousand RISC-V processors on a single TSMC 7nm chip, including: • 1088 energy-efficient ET-Minion 64-bit RISC-V in-order cores each with a vector/tensor unit • 4 high-performance ET-Maxion 64-bit RISC-V out-of-order cores • >160 million bytes of on-chip SRAM • Interfaces for large external memory with low-power LPDDR4x DRAM and eMMC FLASH • PCIe x8 Gen4 and other common I/O interfaces • Innovative low-power architecture and circuit techniques allows entire chip to • Compute at peak rates of 100 to 200 TOPS • Operate using under 20 watts for ML recommendation workloads
ET-SoC-1在台积电7nm芯片上拥有超过1000个RISC-V处理器,包括:•1088年节能ET-Minion 64位RISC-V顺序核每个向量/张量单位•64位高性能ET-Maxion RISC-V无序核心•> 1.6亿字节的片上存储器•大型外部存储器接口与低功耗LPDDR4x DRAM和eMMC FLASH作为PCIe×8•Gen4和其他常见•创新低功耗架构和I / O接口电路技术允许整个芯片•计算在高峰的100年到200年上衣•20瓦毫升的建议下使用操作工作负载
{"title":"Accelerating ML Recommendation with over a Thousand RISC-V/Tensor Processors on Esperanto’s ET-SoC-1 Chip","authors":"D. Ditzel, R. Espasa, Nivard Aymerich, Allen Baum, Tom Berg, Jim Burr, Eric Hao, Jayesh Iyer, Miquel Izquierdo, Shankar Jayaratnam, Darren Jones, Chris Klingner, Jin Kim, Stephen Lee, Marc Lupon, G. Magklis, Bojan Maric, Rajib Nath, Michael Neilly, Duane J. Northcutt, Bill Orner, Jose Renau, Gerard Reves, X. Revés, Tom Riordan, Pedro Sanchez, S. Samudrala, Guillem Sole, Raymond Tang, Tommy Thorn, Francisco Torres, S. Tortella, Daniel Yau","doi":"10.1109/HCS52781.2021.9566904","DOIUrl":"https://doi.org/10.1109/HCS52781.2021.9566904","url":null,"abstract":"The ET-SoC-1 has over a thousand RISC-V processors on a single TSMC 7nm chip, including: • 1088 energy-efficient ET-Minion 64-bit RISC-V in-order cores each with a vector/tensor unit • 4 high-performance ET-Maxion 64-bit RISC-V out-of-order cores • >160 million bytes of on-chip SRAM • Interfaces for large external memory with low-power LPDDR4x DRAM and eMMC FLASH • PCIe x8 Gen4 and other common I/O interfaces • Innovative low-power architecture and circuit techniques allows entire chip to • Compute at peak rates of 100 to 200 TOPS • Operate using under 20 watts for ML recommendation workloads","PeriodicalId":246531,"journal":{"name":"2021 IEEE Hot Chips 33 Symposium (HCS)","volume":"117 5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128376641","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Aquabolt-XL: Samsung HBM2-PIM with in-memory processing for ML accelerators and beyond Aquabolt-XL:三星HBM2-PIM与内存处理ML加速器和超越
Pub Date : 2021-08-22 DOI: 10.1109/HCS52781.2021.9567191
J. Kim, Shinhaeng Kang, Sukhan Lee, Hyeonsu Kim, Woongjae Song, Yuhwan Ro, Seungwon Lee, David Wang, Hyunsung Shin, BengSeng Phuah, Jihyun Choi, J. So, Yeon-Gon Cho, Joonho Song, J. Choi, Jeonghyeon Cho, Kyomin Sohn, Y. Sohn, Kwang-il Park, N. Kim
Using PIM to overcome memory bottleneck • Although various bandwidth increase methods have been proposed, it is physically impossible to achieve a breakthrough increase. - Limited by # of PCB wires, # of CPU ball, and thermal constraints • PIM has been proposed to improve performance of bandwidth-intensive workloads and improve energy efficiency by reducing computing-memory data movement.
利用PIM克服内存瓶颈•虽然提出了各种带宽增加方法,但物理上不可能实现突破性增长。-受PCB线数、CPU球数和热约束的限制•PIM已被提出用于改善带宽密集型工作负载的性能,并通过减少计算内存数据移动来提高能源效率。
{"title":"Aquabolt-XL: Samsung HBM2-PIM with in-memory processing for ML accelerators and beyond","authors":"J. Kim, Shinhaeng Kang, Sukhan Lee, Hyeonsu Kim, Woongjae Song, Yuhwan Ro, Seungwon Lee, David Wang, Hyunsung Shin, BengSeng Phuah, Jihyun Choi, J. So, Yeon-Gon Cho, Joonho Song, J. Choi, Jeonghyeon Cho, Kyomin Sohn, Y. Sohn, Kwang-il Park, N. Kim","doi":"10.1109/HCS52781.2021.9567191","DOIUrl":"https://doi.org/10.1109/HCS52781.2021.9567191","url":null,"abstract":"Using PIM to overcome memory bottleneck • Although various bandwidth increase methods have been proposed, it is physically impossible to achieve a breakthrough increase. - Limited by # of PCB wires, # of CPU ball, and thermal constraints • PIM has been proposed to improve performance of bandwidth-intensive workloads and improve energy efficiency by reducing computing-memory data movement.","PeriodicalId":246531,"journal":{"name":"2021 IEEE Hot Chips 33 Symposium (HCS)","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134201114","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Dynamic Neural Accelerator for Reconfigurable & Energy-efficient Neural Network Inference 可重构节能神经网络推理的动态神经加速器
Pub Date : 2021-08-22 DOI: 10.1109/HCS52781.2021.9566886
Nikolay Nez, Antonio N. Vilchez, H. Zohouri, Oleg Khavin, Sakyasingha Dasgupta
Unique Challenges for AI Inference Hardware at the Edge • Peak TOPS or TOPS/Watt are not ideal measures of performance at the edge. Cannot prioritize performance over power efficiency (throughput/watt) • Many AI Hardware rely on batching to improve utilization. Unsuitable for streaming data (batch size 1) use-case at the edge • AI hardware architectures that fully cache network parameters using large on-chip SRAM cannot be scaled down easily to sizes applicable for edge workloads. • Need adaptability to new workloads and the ability to deploy multiple AI models • AI-specific accelerator needs to operate within heterogenous compute environments • Need for efficient compiler & scheduling to maximize compute utilization • Need for high software robustness and usability
•峰值TOPS或TOPS/Watt不是边缘性能的理想衡量标准。不能优先考虑性能而不是功率效率(吞吐量/瓦特)•许多AI硬件依赖批处理来提高利用率。不适合边缘的流数据(批处理大小为1)用例•使用大型片上SRAM完全缓存网络参数的AI硬件架构不能轻松缩放到适用于边缘工作负载的大小。•需要适应新的工作负载和部署多个AI模型的能力•AI特定加速器需要在异构计算环境中运行•需要高效的编译器和调度以最大限度地提高计算利用率•需要高软件鲁棒性和可用性
{"title":"Dynamic Neural Accelerator for Reconfigurable & Energy-efficient Neural Network Inference","authors":"Nikolay Nez, Antonio N. Vilchez, H. Zohouri, Oleg Khavin, Sakyasingha Dasgupta","doi":"10.1109/HCS52781.2021.9566886","DOIUrl":"https://doi.org/10.1109/HCS52781.2021.9566886","url":null,"abstract":"Unique Challenges for AI Inference Hardware at the Edge • Peak TOPS or TOPS/Watt are not ideal measures of performance at the edge. Cannot prioritize performance over power efficiency (throughput/watt) • Many AI Hardware rely on batching to improve utilization. Unsuitable for streaming data (batch size 1) use-case at the edge • AI hardware architectures that fully cache network parameters using large on-chip SRAM cannot be scaled down easily to sizes applicable for edge workloads. • Need adaptability to new workloads and the ability to deploy multiple AI models • AI-specific accelerator needs to operate within heterogenous compute environments • Need for efficient compiler & scheduling to maximize compute utilization • Need for high software robustness and usability","PeriodicalId":246531,"journal":{"name":"2021 IEEE Hot Chips 33 Symposium (HCS)","volume":"53 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132753106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Intel’s Hyperscale-Ready Infrastructure Processing Unit (IPU) 英特尔的超大规模基础设施处理单元(IPU)
Pub Date : 2021-08-22 DOI: 10.1109/HCS52781.2021.9567455
Bradley Burres, Dan Daly, M. Debbage, Eliel Louzoun, Christine Severns-Williams, Naru Sundar, Nadav Turbovich, Barry Wolford, Yadong Li
Major Advantages of IPUs Separation of Infrastructure & Tenant Guest can fully control the CPU with their SW, while CSP maintains control of the infrastructure and Root of Trust Infrastructure Offload Accelerators help process these task efficiently. Minimize latency and jitter and maximize revenue from CPU Diskless Server Architecture Simplifies data center architecture while adding flexibility for the CSP
ipu的主要优势基础设施和租户客户分离可以完全控制CPU与他们的软件,而CSP保持对基础设施的控制和信任基础设施卸载加速器帮助有效地处理这些任务。最小化延迟和抖动,最大限度地提高CPU无磁盘服务器架构的收益,简化数据中心架构,同时增加CSP的灵活性
{"title":"Intel’s Hyperscale-Ready Infrastructure Processing Unit (IPU)","authors":"Bradley Burres, Dan Daly, M. Debbage, Eliel Louzoun, Christine Severns-Williams, Naru Sundar, Nadav Turbovich, Barry Wolford, Yadong Li","doi":"10.1109/HCS52781.2021.9567455","DOIUrl":"https://doi.org/10.1109/HCS52781.2021.9567455","url":null,"abstract":"Major Advantages of IPUs Separation of Infrastructure & Tenant Guest can fully control the CPU with their SW, while CSP maintains control of the infrastructure and Root of Trust Infrastructure Offload Accelerators help process these task efficiently. Minimize latency and jitter and maximize revenue from CPU Diskless Server Architecture Simplifies data center architecture while adding flexibility for the CSP","PeriodicalId":246531,"journal":{"name":"2021 IEEE Hot Chips 33 Symposium (HCS)","volume":"63 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114708692","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Sapphire Rapids 蓝宝石急流
Pub Date : 2021-08-22 DOI: 10.1109/HCS52781.2021.9566865
Arijit Biswas
Next-Gen Intel Xeon Scalable Processor New Standard for Data Center Architecture Designed for Microservices & AI Workloads Pioneering Advanced Memory & IO Transitions
新一代英特尔至强可扩展处理器为微服务和人工智能工作负载设计的数据中心架构新标准,开创了先进的内存和IO转换
{"title":"Sapphire Rapids","authors":"Arijit Biswas","doi":"10.1109/HCS52781.2021.9566865","DOIUrl":"https://doi.org/10.1109/HCS52781.2021.9566865","url":null,"abstract":"Next-Gen Intel Xeon Scalable Processor New Standard for Data Center Architecture Designed for Microservices & AI Workloads Pioneering Advanced Memory & IO Transitions","PeriodicalId":246531,"journal":{"name":"2021 IEEE Hot Chips 33 Symposium (HCS)","volume":"204 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114754409","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Skydio Autonomy Engine: Enabling The Next Generation Of Autonomous Flight Skydio自主引擎:实现下一代自主飞行
Pub Date : 2021-08-22 DOI: 10.1109/HCS52781.2021.9567400
A. Bachrach
Drones hold the promise of massive positive impact Existing use cases are easier and more reliable New use cases that were previously impossible are enabled
无人机有望带来巨大的积极影响,现有的用例更容易、更可靠,以前不可能实现的新用例也得以实现
{"title":"Skydio Autonomy Engine: Enabling The Next Generation Of Autonomous Flight","authors":"A. Bachrach","doi":"10.1109/HCS52781.2021.9567400","DOIUrl":"https://doi.org/10.1109/HCS52781.2021.9567400","url":null,"abstract":"Drones hold the promise of massive positive impact Existing use cases are easier and more reliable New use cases that were previously impossible are enabled","PeriodicalId":246531,"journal":{"name":"2021 IEEE Hot Chips 33 Symposium (HCS)","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121162119","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
2021 IEEE Hot Chips 33 Symposium (HCS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1