4

I just read that OpenAI's ES uses Adam: "OpenAI’s ES is denoted as “OptimES” (since it uses Adam optimizer)"?? I verified they are correct using the link they posted, (see es_distributed/Optimizers.py). But I don't understand how because the paper says they are using Evolution Strategies as a Scalable Alternative to Reinforcement Learning, which is black box optimization (like most ES's)... So how on earth is Adam used, given this is black box?

nbro
  • 42,615
  • 12
  • 119
  • 217
profPlum
  • 496
  • 2
  • 10

1 Answers1

0

OpenAI's ES updates parameters using learning rate. It may remain constant, decaying or changed in different ways. You can for example use momentum that's commonly used in stochastic gradient descent. So Adam is used to adjust learning rates per parameters.

derjack
  • 116
  • 1
  • 7