For questions about the total variational distance (which can be used as a measure of the distance between probability distributions). See e.g. the paper "On choosing and bounding probability metrics" (2002, by Alison L. Gibbs and Francis Edward Su) for more details about this and other measures.
Questions tagged [total-variational-distance]
3 questions
7
votes
2 answers
Why is KL divergence used so often in Machine Learning?
The KL Divergence is quite easy to compute in closed form for simple distributions -such as Gaussians- but has some not-very-nice properties. For example, it is not symmetrical (thus it is not a metric) and it does not respect the triangular…

Federico Taschin
- 233
- 1
- 6
3
votes
2 answers
When should one prefer using Total Variational Divergence over KL divergence in RL
In RL, both the KL divergence (DKL) and Total variational divergence (DTV) are used to measure the distance between two policies. I'm most familiar with using DKL as an early stopping metric during policy updates to ensure the new policy doesn't…

mugoh
- 531
- 4
- 20
0
votes
0 answers
CS 285 Prof Sergey Levine Lecture, Bounding Derivation for Reinforcement Learning (TRPO)
How can we derive the final result? I can understand the first line, but don't know how the absolute term in the summation is replaced with $2\epsilon…

shashack
- 101
- 3