{"title":"Y-Drop: A Conductance based Dropout for fully connected layers","authors":"Efthymios Georgiou, Georgios Paraskevopoulos, Alexandros Potamianos","doi":"arxiv-2409.09088","DOIUrl":null,"url":null,"abstract":"In this work, we introduce Y-Drop, a regularization method that biases the\ndropout algorithm towards dropping more important neurons with higher\nprobability. The backbone of our approach is neuron conductance, an\ninterpretable measure of neuron importance that calculates the contribution of\neach neuron towards the end-to-end mapping of the network. We investigate the\nimpact of the uniform dropout selection criterion on performance by assigning\nhigher dropout probability to the more important units. We show that forcing\nthe network to solve the task at hand in the absence of its important units\nyields a strong regularization effect. Further analysis indicates that Y-Drop\nyields solutions where more neurons are important, i.e have high conductance,\nand yields robust networks. In our experiments we show that the regularization\neffect of Y-Drop scales better than vanilla dropout w.r.t. the architecture\nsize and consistently yields superior performance over multiple datasets and\narchitecture combinations, with little tuning.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"190 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Neural and Evolutionary Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.09088","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In this work, we introduce Y-Drop, a regularization method that biases the
dropout algorithm towards dropping more important neurons with higher
probability. The backbone of our approach is neuron conductance, an
interpretable measure of neuron importance that calculates the contribution of
each neuron towards the end-to-end mapping of the network. We investigate the
impact of the uniform dropout selection criterion on performance by assigning
higher dropout probability to the more important units. We show that forcing
the network to solve the task at hand in the absence of its important units
yields a strong regularization effect. Further analysis indicates that Y-Drop
yields solutions where more neurons are important, i.e have high conductance,
and yields robust networks. In our experiments we show that the regularization
effect of Y-Drop scales better than vanilla dropout w.r.t. the architecture
size and consistently yields superior performance over multiple datasets and
architecture combinations, with little tuning.