Shreyas Chaudhari;Srinivasa Pranav;José M.F. Moura
{"title":"梯度网络","authors":"Shreyas Chaudhari;Srinivasa Pranav;José M.F. Moura","doi":"10.1109/TSP.2024.3496692","DOIUrl":null,"url":null,"abstract":"Directly parameterizing and learning gradients of functions has widespread significance, with specific applications in inverse problems, generative modeling, and optimal transport. This paper introduces gradient networks (\n<monospace>GradNets</monospace>\n): novel neural network architectures that parameterize gradients of various function classes. \n<monospace>GradNets</monospace>\n exhibit specialized architectural constraints that ensure correspondence to gradient functions. We provide a comprehensive \n<monospace>GradNet</monospace>\n design framework that includes methods for transforming \n<monospace>GradNets</monospace>\n into monotone gradient networks (\n<monospace>mGradNets</monospace>\n), which are guaranteed to represent gradients of convex functions. Our results establish that our proposed \n<monospace>GradNet</monospace>\n (and \n<monospace>mGradNet</monospace>\n) universally approximate the gradients of (convex) functions. Furthermore, these networks can be customized to correspond to specific spaces of potential functions, including transformed sums of (convex) ridge functions. Our analysis leads to two distinct \n<monospace>GradNet</monospace>\n architectures, \n<monospace>GradNet-C</monospace>\n and \n<monospace>GradNet-M</monospace>\n, and we describe the corresponding monotone versions, \n<monospace>mGradNet-C</monospace>\n and \n<monospace>mGradNet-M</monospace>\n. Our empirical results demonstrate that these architectures provide efficient parameterizations and outperform existing methods by up to 15 dB in gradient field tasks and by up to 11 dB in Hamiltonian dynamics learning tasks.","PeriodicalId":13330,"journal":{"name":"IEEE Transactions on Signal Processing","volume":"73 ","pages":"324-339"},"PeriodicalIF":4.6000,"publicationDate":"2024-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Gradient Networks\",\"authors\":\"Shreyas Chaudhari;Srinivasa Pranav;José M.F. Moura\",\"doi\":\"10.1109/TSP.2024.3496692\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Directly parameterizing and learning gradients of functions has widespread significance, with specific applications in inverse problems, generative modeling, and optimal transport. This paper introduces gradient networks (\\n<monospace>GradNets</monospace>\\n): novel neural network architectures that parameterize gradients of various function classes. \\n<monospace>GradNets</monospace>\\n exhibit specialized architectural constraints that ensure correspondence to gradient functions. We provide a comprehensive \\n<monospace>GradNet</monospace>\\n design framework that includes methods for transforming \\n<monospace>GradNets</monospace>\\n into monotone gradient networks (\\n<monospace>mGradNets</monospace>\\n), which are guaranteed to represent gradients of convex functions. Our results establish that our proposed \\n<monospace>GradNet</monospace>\\n (and \\n<monospace>mGradNet</monospace>\\n) universally approximate the gradients of (convex) functions. Furthermore, these networks can be customized to correspond to specific spaces of potential functions, including transformed sums of (convex) ridge functions. Our analysis leads to two distinct \\n<monospace>GradNet</monospace>\\n architectures, \\n<monospace>GradNet-C</monospace>\\n and \\n<monospace>GradNet-M</monospace>\\n, and we describe the corresponding monotone versions, \\n<monospace>mGradNet-C</monospace>\\n and \\n<monospace>mGradNet-M</monospace>\\n. Our empirical results demonstrate that these architectures provide efficient parameterizations and outperform existing methods by up to 15 dB in gradient field tasks and by up to 11 dB in Hamiltonian dynamics learning tasks.\",\"PeriodicalId\":13330,\"journal\":{\"name\":\"IEEE Transactions on Signal Processing\",\"volume\":\"73 \",\"pages\":\"324-339\"},\"PeriodicalIF\":4.6000,\"publicationDate\":\"2024-11-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Signal Processing\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10752831/\",\"RegionNum\":2,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Signal Processing","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10752831/","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
Directly parameterizing and learning gradients of functions has widespread significance, with specific applications in inverse problems, generative modeling, and optimal transport. This paper introduces gradient networks (
GradNets
): novel neural network architectures that parameterize gradients of various function classes.
GradNets
exhibit specialized architectural constraints that ensure correspondence to gradient functions. We provide a comprehensive
GradNet
design framework that includes methods for transforming
GradNets
into monotone gradient networks (
mGradNets
), which are guaranteed to represent gradients of convex functions. Our results establish that our proposed
GradNet
(and
mGradNet
) universally approximate the gradients of (convex) functions. Furthermore, these networks can be customized to correspond to specific spaces of potential functions, including transformed sums of (convex) ridge functions. Our analysis leads to two distinct
GradNet
architectures,
GradNet-C
and
GradNet-M
, and we describe the corresponding monotone versions,
mGradNet-C
and
mGradNet-M
. Our empirical results demonstrate that these architectures provide efficient parameterizations and outperform existing methods by up to 15 dB in gradient field tasks and by up to 11 dB in Hamiltonian dynamics learning tasks.
期刊介绍:
The IEEE Transactions on Signal Processing covers novel theory, algorithms, performance analyses and applications of techniques for the processing, understanding, learning, retrieval, mining, and extraction of information from signals. The term “signal” includes, among others, audio, video, speech, image, communication, geophysical, sonar, radar, medical and musical signals. Examples of topics of interest include, but are not limited to, information processing and the theory and application of filtering, coding, transmitting, estimating, detecting, analyzing, recognizing, synthesizing, recording, and reproducing signals.