In Bayesian inference prior hyperparameters are chosen subjectively or estimated using empirical Bayes methods. Generalised Bayesian Inference (GBI) also has a learning rate hyperparameter. This is compounded in Semi-Modular Inference (SMI), a GBI framework for multiple datasets (multi-modular problems). As part of any GBI workflow it is necessary to check sensitivity to the choice of hyperparameters, but running MCMC or fitting a variational approximation at each of the hyperparameter values of interest is impractical. Simulation-based Inference has been used by previous authors to amortise over data and hyperparameters, fitting a posterior approximation targeting the forward-KL divergence. However, for GBI and SMI posteriors, it is not possible to amortise over data, as there is no generative model. Working with a variational family parameterised by a conditional normalising flow, we give a direct variational approximation for GBI and SMI posteriors, targeting the reverse-KL divergence, and amortised over prior and loss hyperparameters at fixed data. This can be sampled efficiently at different hyperparameter values without refitting, and supports efficient robustness checks and hyperparameter selection. We show that there exist amortised conditional normalising-flow architectures which are universal approximators. We illustrate our methods with an epidemiological example well known in SMI work and then give the motivating application, a spatial location-prediction task for linguistic-profile data. SMI gives improved prediction with hyperparameters chosen using our amortised framework. The code is available online.
翻译:暂无翻译