Mixtures of experts have become an indispensable tool for flexible modelling in a supervised learning context, and sparse Gaussian processes (GP) have shown promise as a leading candidate for the experts in such models. In this article, we propose to design the gating network for selecting the experts from such mixtures of sparse GPs using a deep neural network (DNN). Furthermore, an ultra-fast one pass algorithm called Cluster-Classify-Regress (CCR) is leveraged to approximate the maximum a posteriori (MAP) estimator extremely quickly. This powerful combination of model and algorithm together delivers a novel method which is flexible, robust, and extremely efficient. In particular, the method is able to outperform competing methods in terms of accuracy and uncertainty quantification. The cost is competitive on low-dimensional and small data sets but is significantly lower for higher dimensional and big data sets. Iteratively maximizing the distribution of experts given allocations and allocations given experts does not provide significant improvement, which indicates that the algorithm achieves a good approximation to the local MAP estimator very fast. This insight can be useful also in the context of other mixture of experts models.
翻译:专家混合已成为在有监督的学习环境中进行灵活建模的不可或缺的工具,而稀有的高斯进程(GP)作为这类模型专家的主要候选人表现出了希望。在本条中,我们提议设计一个格子网络,利用深神经网络从这种稀多的GP混合物中挑选专家。此外,超快的一种传算法,称为Croup-Clasti-Regrest(CCR),被用来极快地接近事后估计器(MAP)的最大比例。这种强大的模型和算法组合提供了一种灵活、有力和极有效率的新方法。特别是,该方法能够在准确性和不确定性的量化方面超越相互竞争的方法。成本在低维和小数据集上是竞争性的,但对于高维和大数据集而言则要低得多。 尽量扩大专家的分配和分配并不能提供显著的改进,这说明该算法能够非常迅速地接近当地MAP估计器。这一洞察器在其他专家的混合模型中也很有用。