EWGN: Elastic weight generation and context switching in deep learning

Show simple item record

dc.contributor.author Sawant, Shriraj P.
dc.contributor.author Miyapuram, Krishna Prasad
dc.coverage.spatial United States of America
dc.date.accessioned 2025-06-12T06:23:42Z
dc.date.available 2025-06-12T06:23:42Z
dc.date.issued 2025-06
dc.identifier.citation Sawant, Shriraj P. and Miyapuram, Krishna Prasad, "EWGN: Elastic weight generation and context switching in deep learning", arXiv, Cornell University Library, DOI: arXiv:2506.02065, Jun. 2025.
dc.identifier.uri https://doi.org/10.48550/arXiv.2506.02065
dc.identifier.uri https://repository.iitgn.ac.in/handle/123456789/11526
dc.description.abstract The ability to learn and retain a wide variety of tasks is a hallmark of human intelligence that has inspired research in artificial general intelligence. Continual learning approaches provide a significant step towards achieving this goal. It has been known that task variability and context switching are challenging for learning in neural networks. Catastrophic forgetting refers to the poor performance on retention of a previously learned task when a new task is being learned. Switching between different task contexts can be a useful approach to mitigate the same by preventing the interference between the varying task weights of the network. This paper introduces Elastic Weight Generative Networks (EWGN) as an idea for context switching between two different tasks. The proposed EWGN architecture uses an additional network that generates the weights of the primary network dynamically while consolidating the weights learned. The weight generation is input-dependent and thus enables context switching. Using standard computer vision datasets, namely MNIST and fashion-MNIST, we analyse the retention of previously learned task representations in Fully Connected Networks, Convolutional Neural Networks, and EWGN architectures with Stochastic Gradient Descent and Elastic Weight Consolidation learning algorithms. Understanding dynamic weight generation and context-switching ability can be useful in enabling continual learning for improved performance.
dc.description.statementofresponsibility by Shriraj P. Sawant and Krishna Prasad Miyapuram
dc.language.iso en_US
dc.publisher Cornell University Library
dc.title EWGN: Elastic weight generation and context switching in deep learning
dc.type Article
dc.relation.journal arXiv


Files in this item

Files Size Format View

There are no files associated with this item.

This item appears in the following Collection(s)

Show simple item record

Search Digital Repository


Browse

My Account