Questions tagged [total-variational-distance]

For questions about the total variational distance (which can be used as a measure of the distance between probability distributions). See e.g. the paper "On choosing and bounding probability metrics" (2002, by Alison L. Gibbs and Francis Edward Su) for more details about this and other measures.

2 questions
8
votes
2 answers

Why is KL divergence used so often in Machine Learning?

The KL Divergence is quite easy to compute in closed form for simple distributions -such as Gaussians- but has some not-very-nice properties. For example, it is not symmetrical (thus it is not a metric) and it does not respect the triangular…
4
votes
2 answers

When should one prefer using Total Variational Divergence over KL divergence in RL

In RL, both the KL divergence (DKL) and Total variational divergence (DTV) are used to measure the distance between two policies. I'm most familiar with using DKL as an early stopping metric during policy updates to ensure the new policy doesn't…