{"title":"自对焦前向算法","authors":"Xing Chen, Dongshu Liu, Jeremie Laydevant, Julie Grollier","doi":"arxiv-2409.11593","DOIUrl":null,"url":null,"abstract":"The Forward-Forward (FF) algorithm is a recent, purely forward-mode learning\nmethod, that updates weights locally and layer-wise and supports supervised as\nwell as unsupervised learning. These features make it ideal for applications\nsuch as brain-inspired learning, low-power hardware neural networks, and\ndistributed learning in large models. However, while FF has shown promise on\nwritten digit recognition tasks, its performance on natural images and\ntime-series remains a challenge. A key limitation is the need to generate\nhigh-quality negative examples for contrastive learning, especially in\nunsupervised tasks, where versatile solutions are currently lacking. To address\nthis, we introduce the Self-Contrastive Forward-Forward (SCFF) method, inspired\nby self-supervised contrastive learning. SCFF generates positive and negative\nexamples applicable across different datasets, surpassing existing local\nforward algorithms for unsupervised classification accuracy on MNIST (MLP:\n98.7%), CIFAR-10 (CNN: 80.75%), and STL-10 (CNN: 77.3%). Additionally, SCFF is\nthe first to enable FF training of recurrent neural networks, opening the door\nto more complex tasks and continuous-time video and text processing.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"18 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Self-Contrastive Forward-Forward Algorithm\",\"authors\":\"Xing Chen, Dongshu Liu, Jeremie Laydevant, Julie Grollier\",\"doi\":\"arxiv-2409.11593\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The Forward-Forward (FF) algorithm is a recent, purely forward-mode learning\\nmethod, that updates weights locally and layer-wise and supports supervised as\\nwell as unsupervised learning. These features make it ideal for applications\\nsuch as brain-inspired learning, low-power hardware neural networks, and\\ndistributed learning in large models. However, while FF has shown promise on\\nwritten digit recognition tasks, its performance on natural images and\\ntime-series remains a challenge. A key limitation is the need to generate\\nhigh-quality negative examples for contrastive learning, especially in\\nunsupervised tasks, where versatile solutions are currently lacking. To address\\nthis, we introduce the Self-Contrastive Forward-Forward (SCFF) method, inspired\\nby self-supervised contrastive learning. SCFF generates positive and negative\\nexamples applicable across different datasets, surpassing existing local\\nforward algorithms for unsupervised classification accuracy on MNIST (MLP:\\n98.7%), CIFAR-10 (CNN: 80.75%), and STL-10 (CNN: 77.3%). Additionally, SCFF is\\nthe first to enable FF training of recurrent neural networks, opening the door\\nto more complex tasks and continuous-time video and text processing.\",\"PeriodicalId\":501347,\"journal\":{\"name\":\"arXiv - CS - Neural and Evolutionary Computing\",\"volume\":\"18 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Neural and Evolutionary Computing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.11593\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Neural and Evolutionary Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.11593","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
The Forward-Forward (FF) algorithm is a recent, purely forward-mode learning
method, that updates weights locally and layer-wise and supports supervised as
well as unsupervised learning. These features make it ideal for applications
such as brain-inspired learning, low-power hardware neural networks, and
distributed learning in large models. However, while FF has shown promise on
written digit recognition tasks, its performance on natural images and
time-series remains a challenge. A key limitation is the need to generate
high-quality negative examples for contrastive learning, especially in
unsupervised tasks, where versatile solutions are currently lacking. To address
this, we introduce the Self-Contrastive Forward-Forward (SCFF) method, inspired
by self-supervised contrastive learning. SCFF generates positive and negative
examples applicable across different datasets, surpassing existing local
forward algorithms for unsupervised classification accuracy on MNIST (MLP:
98.7%), CIFAR-10 (CNN: 80.75%), and STL-10 (CNN: 77.3%). Additionally, SCFF is
the first to enable FF training of recurrent neural networks, opening the door
to more complex tasks and continuous-time video and text processing.