5

I am running a basic DQN (Deep Q-Network) on the Pong environment. Not a CNN, just a 3 layer linear neural net with ReLUs.

It seems to work for the most part, but at some point, my model suffers from catastrophic performance loss:

Catastrophic forgetting

  1. What is really the reason for that?

  2. What are the common ways to avoid this? Clipping the gradients? What else?

(Reloading from previous successful checkpoints feels more like a hack, rather than a proper solution to this issue.)

nbro
  • 42,615
  • 12
  • 119
  • 217
Muppet
  • 151
  • 3

0 Answers0