Bayesian methods have shown success in deep learning applications. For example, in predictive tasks, Bayesian neural networks leverage Bayesian reasoning of model uncertainty to improve the reliability and uncertainty awareness of deep neural networks. In generative modeling domain, many widely used deep generative models, such as deep latent variable models, require approximate Bayesian inference to infer their latent variables for the training. In this chapter, we provide an introduction to approximate inference techniques as Bayesian computation methods applied to deep learning models, with a focus on Bayesian neural networks and deep generative models. We review two arguably most popular approximate Bayesian computational methods, stochastic gradient Markov chain Monte Carlo (SG-MCMC) and variational inference (VI), and explain their unique challenges in posterior inference as well as the solutions when applied to deep learning models.
翻译:贝叶斯方法在深度学习应用中已展现出显著成效。例如,在预测任务中,贝叶斯神经网络利用模型不确定性的贝叶斯推理,提升了深度神经网络的可靠性与不确定性感知能力。在生成建模领域,许多广泛使用的深度生成模型(如深度隐变量模型)需借助近似贝叶斯推断来训练其隐变量。本章将系统介绍应用于深度学习模型的近似推断技术——即贝叶斯计算方法,重点关注贝叶斯神经网络与深度生成模型。我们回顾了当前两种最主流的近似贝叶斯计算方法:随机梯度马尔可夫链蒙特卡罗(SG-MCMC)与变分推断(VI),阐释它们在深度学习中后验推断面临的特殊挑战及相应解决方案。