{"title":"Symmetry Breaking in Neural Network Optimization: Insights from Input Dimension Expansion","authors":"Jun-Jie Zhang, Nan Cheng, Fu-Peng Li, Xiu-Cheng Wang, Jian-Nan Chen, Long-Gang Pang, Deyu Meng","doi":"arxiv-2409.06402","DOIUrl":null,"url":null,"abstract":"Understanding the mechanisms behind neural network optimization is crucial\nfor improving network design and performance. While various optimization\ntechniques have been developed, a comprehensive understanding of the underlying\nprinciples that govern these techniques remains elusive. Specifically, the role\nof symmetry breaking, a fundamental concept in physics, has not been fully\nexplored in neural network optimization. This gap in knowledge limits our\nability to design networks that are both efficient and effective. Here, we\npropose the symmetry breaking hypothesis to elucidate the significance of\nsymmetry breaking in enhancing neural network optimization. We demonstrate that\na simple input expansion can significantly improve network performance across\nvarious tasks, and we show that this improvement can be attributed to the\nunderlying symmetry breaking mechanism. We further develop a metric to quantify\nthe degree of symmetry breaking in neural networks, providing a practical\napproach to evaluate and guide network design. Our findings confirm that\nsymmetry breaking is a fundamental principle that underpins various\noptimization techniques, including dropout, batch normalization, and\nequivariance. By quantifying the degree of symmetry breaking, our work offers a\npractical technique for performance enhancement and a metric to guide network\ndesign without the need for complete datasets and extensive training processes.","PeriodicalId":501312,"journal":{"name":"arXiv - MATH - Mathematical Physics","volume":"23 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - MATH - Mathematical Physics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.06402","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Understanding the mechanisms behind neural network optimization is crucial
for improving network design and performance. While various optimization
techniques have been developed, a comprehensive understanding of the underlying
principles that govern these techniques remains elusive. Specifically, the role
of symmetry breaking, a fundamental concept in physics, has not been fully
explored in neural network optimization. This gap in knowledge limits our
ability to design networks that are both efficient and effective. Here, we
propose the symmetry breaking hypothesis to elucidate the significance of
symmetry breaking in enhancing neural network optimization. We demonstrate that
a simple input expansion can significantly improve network performance across
various tasks, and we show that this improvement can be attributed to the
underlying symmetry breaking mechanism. We further develop a metric to quantify
the degree of symmetry breaking in neural networks, providing a practical
approach to evaluate and guide network design. Our findings confirm that
symmetry breaking is a fundamental principle that underpins various
optimization techniques, including dropout, batch normalization, and
equivariance. By quantifying the degree of symmetry breaking, our work offers a
practical technique for performance enhancement and a metric to guide network
design without the need for complete datasets and extensive training processes.