Weight clipping for deep continual and reinforcement learning
Many failures in deep continual and reinforcement learning are associated with increasing
magnitudes of the weights, making them hard to change and potentially causing overfitting …
magnitudes of the weights, making them hard to change and potentially causing overfitting …
Addressing loss of plasticity and catastrophic forgetting in continual learning
M Elsayed, AR Mahmood - arXiv preprint arXiv:2404.00781, 2024 - arxiv.org
Deep representation learning methods struggle with continual learning, suffering from both
catastrophic forgetting of useful units and loss of plasticity, often due to rigid and unuseful …
catastrophic forgetting of useful units and loss of plasticity, often due to rigid and unuseful …
Improving deep reinforcement learning by reducing the chain effect of value and policy churn
Deep neural networks provide Reinforcement Learning (RL) powerful function
approximators to address large-scale decision-making problems. However, these …
approximators to address large-scale decision-making problems. However, these …
Deep Policy Gradient Methods Without Batch Updates, Target Networks, or Replay Buffers
Modern deep policy gradient methods achieve effective performance on simulated robotic
tasks, but they all require large replay buffers or expensive batch updates, or both, making …
tasks, but they all require large replay buffers or expensive batch updates, or both, making …
Elephant neural networks: Born to be a continual learner
Q Lan, AR Mahmood - arXiv preprint arXiv:2310.01365, 2023 - arxiv.org
Catastrophic forgetting remains a significant challenge to continual learning for decades.
While recent works have proposed effective methods to mitigate this problem, they mainly …
While recent works have proposed effective methods to mitigate this problem, they mainly …