For questions about the re-parameterization trick, which is used e.g. in VAEs.
Questions tagged [reparameterization-trick]
3 questions
4
votes
3 answers
In the VAE, why is $z \sim \mathcal{N}(\mu, \sigma^2)$ equivalent to $z = \mu + \sigma \odot \epsilon$?
In the reparameterization trick of a Variational Autoencoder (VAE), instead of sampling noise $z$ from $z \sim \mathcal{N}(\mu, \sigma^2)$, we can use a different method: $z = \mu + \sigma \odot \epsilon$, where $\epsilon \sim \mathcal{N}(0,1)$. I'm…
user77925
2
votes
2 answers
Is the re-parameterization trick necessary in the policy gradient method?
If we want to learn a stochastic policy with the policy gradient method, we have to sample from the distribution to get an action.
Wouldn't this lead to the same issue that variational autoencoders face without the reparameterization trick, where…
Sam
- 205
- 1
- 5
0
votes
1 answer
Why clamp std for reparameterization trick between -20 and 2?
In the Soft Actor Critic Paper (found here https://arxiv.org/pdf/1801.01290.pdf), they use a neural network to approximate a diagonal gaussian distribution. In the sample function you can see that it has a function called reparameterize. As you can…
chadmc
- 13
- 3