On Fri 9 Jun an 11am we'll be discussing "Likelihood-free inference via classification" by Gutmann, et al. (https://arxiv.org/pdf/1407.4981.pdf , abstract below) in the Z6 fishbowl. It should be an interesting discussion for anyone interested in inference of intractable generative models and approximate Bayesian inference in general, so if you can make it, be sure to come!
Increasingly complex generative models are being used across disciplines as they allow for realistic characterization of data, but a common difficulty with them is the prohibitively large computational cost to evaluate the likelihood function and thus to perform likelihood-based statistical inference. A likelihood-free inference framework has emerged where the parameters are identified by finding values that yield simulated data resembling the observed data. While widely applicable, a major difficulty in this framework is how to measure the discrepancy between the simulated and observed data. Transforming the original problem into a problem of classifying the data into simulated versus observed, we find that classification accuracy can be used to assess the discrepancy. The complete arsenal of classification methods becomes thereby available for inference of intractable generative models. We validate our approach using theory and simulations for both point estimation and Bayesian inference, and demonstrate its use on real data by inferring an individual-based epidemiological model for bacterial infections in child care centers.