Metadata only
Date
2024-07Type
- Conference Paper
ETH Bibliography
yes
Altmetrics
Abstract
Online learning holds the promise of enabling efficient long-term credit assignment in recurrent neural networks. However, current algorithms fall short of offline backpropagation by either not being scalable or failing to learn long-range dependencies. Here we present a high-performance online learning algorithm that merely doubles the memory and computational requirements of a single inference pass. We achieve this by leveraging independent recurrent modules in multi-layer networks, an architectural motif that has recently been shown to be particularly powerful. Experiments on synthetic memory problems and on the challenging long-range arena benchmark suite reveal that our algorithm performs competitively, establishing a new standard for what can be achieved through online learning. This ability to learn long-range dependencies offers a new perspective on learning in the brain and opens a promising avenue in neuromorphic computing. Show more
Publication status
publishedExternal links
Book title
Advances in Neural Information Processing Systems 36Pages / Article No.
Publisher
CurranEvent
Subject
Machine Learning (cs.LG); Neural and Evolutionary Computing (cs.NE); FOS: Computer and information sciencesOrganisational unit
03672 - Steger, Angelika / Steger, Angelika
Related publications and datasets
Is new version of: https://openreview.net/forum?id=Wa1GGPqjUn
Notes
Poster presented on December 12, 2023.More
Show all metadata
ETH Bibliography
yes
Altmetrics