variational inference (VI) is a popular approach in Bayesian inference, that
looks for the best approximation of the posterior distribution within a
parametric family, minimizing a loss that is typically the (reverse)
Kullback-Leibler (KL) divergence. Despite its empirical success, the