Algorithm xxx: Efficient algorithms for computing a rank-revealing UTV factorization on parallel computing architectures

IF 2.7 1区 数学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING ACM Transactions on Mathematical Software Pub Date : 2022-03-25 DOI:10.1145/3507466
N. Heavner, F. D. Igual, G. Quintana-Ortí, P.G. Martinsson
{"title":"Algorithm xxx: Efficient algorithms for computing a rank-revealing UTV factorization on parallel computing architectures","authors":"N. Heavner, F. D. Igual, G. Quintana-Ortí, P.G. Martinsson","doi":"10.1145/3507466","DOIUrl":null,"url":null,"abstract":"\n The randomized singular value decomposition (RSVD) is by now a well established technique for efficiently computing an approximate singular value decomposition of a matrix. Building on the ideas that underpin the RSVD, the recently proposed algorithm “randUTV” computes a\n full\n factorization of a given matrix that provides low-rank approximations with near-optimal error. Because the bulk of\n randUTV\n is cast in terms of communication-efficient operations like matrix-matrix multiplication and unpivoted QR factorizations, it is faster than competing rank-revealing factorization methods like column-pivoted QR in most high performance computational settings. In this article, optimized\n randUTV\n implementations are presented for both shared-memory and distributed-memory computing environments. For shared memory,\n randUTV\n is redesigned in terms of an\n algorithm-by-blocks\n that, together with a runtime task scheduler, eliminates bottlenecks from data synchronization points to achieve acceleration over the standard\n blocked algorithm\n , based on a purely fork-join approach. The distributed-memory implementation is based on the ScaLAPACK library. The performances of our new codes compare favorably with competing factorizations available on both shared-memory and distributed-memory architectures.\n","PeriodicalId":50935,"journal":{"name":"ACM Transactions on Mathematical Software","volume":" ","pages":""},"PeriodicalIF":2.7000,"publicationDate":"2022-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ACM Transactions on Mathematical Software","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1145/3507466","RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, SOFTWARE ENGINEERING","Score":null,"Total":0}
引用次数: 2

Abstract

The randomized singular value decomposition (RSVD) is by now a well established technique for efficiently computing an approximate singular value decomposition of a matrix. Building on the ideas that underpin the RSVD, the recently proposed algorithm “randUTV” computes a full factorization of a given matrix that provides low-rank approximations with near-optimal error. Because the bulk of randUTV is cast in terms of communication-efficient operations like matrix-matrix multiplication and unpivoted QR factorizations, it is faster than competing rank-revealing factorization methods like column-pivoted QR in most high performance computational settings. In this article, optimized randUTV implementations are presented for both shared-memory and distributed-memory computing environments. For shared memory, randUTV is redesigned in terms of an algorithm-by-blocks that, together with a runtime task scheduler, eliminates bottlenecks from data synchronization points to achieve acceleration over the standard blocked algorithm , based on a purely fork-join approach. The distributed-memory implementation is based on the ScaLAPACK library. The performances of our new codes compare favorably with competing factorizations available on both shared-memory and distributed-memory architectures.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
算法xxx:在并行计算架构上计算显示秩的UTV分解的有效算法
随机奇异值分解(RSVD)是目前公认的有效计算矩阵近似奇异值分解的技术。基于支持RSVD的思想,最近提出的算法“randUTV”计算给定矩阵的全因子分解,该矩阵提供具有接近最优误差的低秩近似。由于randUTV的大部分是根据矩阵矩阵乘法和非分解QR因子分解等通信高效运算进行的,因此在大多数高性能计算环境中,它比列轴QR等竞争性的秩揭示因子分解方法更快。在本文中,为共享内存和分布式内存计算环境提供了优化的randUTV实现。对于共享内存,randUTV根据块算法进行了重新设计,该算法与运行时任务调度器一起,消除了数据同步点的瓶颈,以实现比标准块算法更快的速度,基于纯粹的fork-join方法。分布式内存实现基于ScaLAPACK库。我们的新代码的性能与共享内存和分布式内存架构上的竞争因子分解相比是有利的。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
ACM Transactions on Mathematical Software
ACM Transactions on Mathematical Software 工程技术-计算机:软件工程
CiteScore
5.00
自引率
3.70%
发文量
50
审稿时长
>12 weeks
期刊介绍: As a scientific journal, ACM Transactions on Mathematical Software (TOMS) documents the theoretical underpinnings of numeric, symbolic, algebraic, and geometric computing applications. It focuses on analysis and construction of algorithms and programs, and the interaction of programs and architecture. Algorithms documented in TOMS are available as the Collected Algorithms of the ACM at calgo.acm.org.
期刊最新文献
Algorithm xxx: A Covariate-Dependent Approach to Gaussian Graphical Modeling in R Remark on Algorithm 1012: Computing projections with large data sets PyOED: An Extensible Suite for Data Assimilation and Model-Constrained Optimal Design of Experiments Avoiding breakdown in incomplete factorizations in low precision arithmetic Algorithm xxx: PyGenStability, a multiscale community detection with generalized Markov Stability
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1