Skip to content

Deep RL faces the problem of plasticity loss, which is the loss of ability of neural networks to adapt to new data/conditions. We discover that regularizing singular values is a promising direction towards mitigating neural network plasticity.

Notifications You must be signed in to change notification settings

arvindrajaraman/plasticity-rl

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

30 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Investigating Plasticity in RL

By: Verona Teo*, Arvind Rajaraman*, and Seyone Chithrananda*

* Equal contribution and co-authorship

This repo explores the problem of plasticity loss in deep reinforcement learning. Plasticity loss is the phenomenon of neural networks losing their ability to classify/regress on new data, when trained on old data until convergence. This term is derived from neuroplasticity loss, which is the loss of ability of human brains adapting to new experiences over time (which explains why it's harder for adults to learn a language as compared to children).

image image

Specifically, we investigate the loss of plasticity in deep Q-networks and analyze how the weights, rank, and activation neurons change over time in several experiments. We explore:

  • Successive regularization for later layers in a neural network
  • Regularizing singular values to create desirable convergence properties
  • Selectively regularizing only during certain time steps
  • Regularizing weights towards the distribution of initial weights, not just towards 0 weight magnitude

This project was done as a part of CS 285 (Deep Reinforcement Learning) at UC Berkeley in the Fall 2023 semester.

About

Deep RL faces the problem of plasticity loss, which is the loss of ability of neural networks to adapt to new data/conditions. We discover that regularizing singular values is a promising direction towards mitigating neural network plasticity.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •  

Languages