That’s the title of a recent article by Yuling Yao, Aki Vehtari, Daniel Simpson, and myself, which presents some diagnostics for variational approximations to posterior inference:

We were motivated to write this paper by the success/failure of ADVI, the automatic variational inference algorithm devised by Alp Kucukelbir et al. The success was that ADVI solved some pretty big problems very fast; the failure was that in lots of examples, ADVI gave really bad answers. We’re still working on figuring all this out—a big issue seems to be scaling, as some of the most horrible disasters are occurring when parameters are far from unit scale. One issue here is that it’s easier to come up with a method than to get it to work reliably, another seems to be a problem in academic incentives, where there’s more of a motivation to attack new problems than to get old things working.

In any case, on the way to trying to fix ADVI and develop a practical automatic method, we thought it could make sense to formalize some of our approaches for checking how well it’s working, as there are various ways that a variational solution, or estimate, can be compared to the full objective function (in Bayesian terms, the log-posterior density).

For an outside perspective on this work, here’s a post by Stephanie Hyland suggesting some directions for future research.

The post Yes, but did it work? Evaluating variational inference appeared first on Statistical Modeling, Causal Inference, and Social Science.