Jakub Vrabel, Ori Shem-Ur, Yaron Oz, David Krueger
{"title":"深度神经网络的输入空间模式连接性","authors":"Jakub Vrabel, Ori Shem-Ur, Yaron Oz, David Krueger","doi":"arxiv-2409.05800","DOIUrl":null,"url":null,"abstract":"We extend the concept of loss landscape mode connectivity to the input space\nof deep neural networks. Mode connectivity was originally studied within\nparameter space, where it describes the existence of low-loss paths between\ndifferent solutions (loss minimizers) obtained through gradient descent. We\npresent theoretical and empirical evidence of its presence in the input space\nof deep networks, thereby highlighting the broader nature of the phenomenon. We\nobserve that different input images with similar predictions are generally\nconnected, and for trained models, the path tends to be simple, with only a\nsmall deviation from being a linear path. Our methodology utilizes real,\ninterpolated, and synthetic inputs created using the input optimization\ntechnique for feature visualization. We conjecture that input space mode\nconnectivity in high-dimensional spaces is a geometric effect that takes place\neven in untrained models and can be explained through percolation theory. We\nexploit mode connectivity to obtain new insights about adversarial examples and\ndemonstrate its potential for adversarial detection. Additionally, we discuss\napplications for the interpretability of deep networks.","PeriodicalId":501520,"journal":{"name":"arXiv - PHYS - Statistical Mechanics","volume":"181 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Input Space Mode Connectivity in Deep Neural Networks\",\"authors\":\"Jakub Vrabel, Ori Shem-Ur, Yaron Oz, David Krueger\",\"doi\":\"arxiv-2409.05800\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"We extend the concept of loss landscape mode connectivity to the input space\\nof deep neural networks. Mode connectivity was originally studied within\\nparameter space, where it describes the existence of low-loss paths between\\ndifferent solutions (loss minimizers) obtained through gradient descent. We\\npresent theoretical and empirical evidence of its presence in the input space\\nof deep networks, thereby highlighting the broader nature of the phenomenon. We\\nobserve that different input images with similar predictions are generally\\nconnected, and for trained models, the path tends to be simple, with only a\\nsmall deviation from being a linear path. Our methodology utilizes real,\\ninterpolated, and synthetic inputs created using the input optimization\\ntechnique for feature visualization. We conjecture that input space mode\\nconnectivity in high-dimensional spaces is a geometric effect that takes place\\neven in untrained models and can be explained through percolation theory. We\\nexploit mode connectivity to obtain new insights about adversarial examples and\\ndemonstrate its potential for adversarial detection. Additionally, we discuss\\napplications for the interpretability of deep networks.\",\"PeriodicalId\":501520,\"journal\":{\"name\":\"arXiv - PHYS - Statistical Mechanics\",\"volume\":\"181 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - PHYS - Statistical Mechanics\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.05800\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - PHYS - Statistical Mechanics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.05800","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Input Space Mode Connectivity in Deep Neural Networks
We extend the concept of loss landscape mode connectivity to the input space
of deep neural networks. Mode connectivity was originally studied within
parameter space, where it describes the existence of low-loss paths between
different solutions (loss minimizers) obtained through gradient descent. We
present theoretical and empirical evidence of its presence in the input space
of deep networks, thereby highlighting the broader nature of the phenomenon. We
observe that different input images with similar predictions are generally
connected, and for trained models, the path tends to be simple, with only a
small deviation from being a linear path. Our methodology utilizes real,
interpolated, and synthetic inputs created using the input optimization
technique for feature visualization. We conjecture that input space mode
connectivity in high-dimensional spaces is a geometric effect that takes place
even in untrained models and can be explained through percolation theory. We
exploit mode connectivity to obtain new insights about adversarial examples and
demonstrate its potential for adversarial detection. Additionally, we discuss
applications for the interpretability of deep networks.