Weight clipping for deep continual and reinforcement learning

M Elsayed, Q Lan, C Lyle, AR Mahmood - arXiv preprint arXiv:2407.01704, 2024 - arxiv.org
Many failures in deep continual and reinforcement learning are associated with increasing
magnitudes of the weights, making them hard to change and potentially causing overfitting …

Addressing loss of plasticity and catastrophic forgetting in continual learning

M Elsayed, AR Mahmood - arXiv preprint arXiv:2404.00781, 2024 - arxiv.org
Deep representation learning methods struggle with continual learning, suffering from both
catastrophic forgetting of useful units and loss of plasticity, often due to rigid and unuseful …

Improving deep reinforcement learning by reducing the chain effect of value and policy churn

H Tang, G Berseth - arXiv preprint arXiv:2409.04792, 2024 - arxiv.org
Deep neural networks provide Reinforcement Learning (RL) powerful function
approximators to address large-scale decision-making problems. However, these …

Deep Policy Gradient Methods Without Batch Updates, Target Networks, or Replay Buffers

G Vasan, M Elsayed, A Azimi, J He, F Shariar… - arXiv preprint arXiv …, 2024 - arxiv.org
Modern deep policy gradient methods achieve effective performance on simulated robotic
tasks, but they all require large replay buffers or expensive batch updates, or both, making …

Elephant neural networks: Born to be a continual learner

Q Lan, AR Mahmood - arXiv preprint arXiv:2310.01365, 2023 - arxiv.org
Catastrophic forgetting remains a significant challenge to continual learning for decades.
While recent works have proposed effective methods to mitigate this problem, they mainly …