Distributional shift is one of the major obstacles when transferring machine learning prediction systems from the lab to the real world. To tackle this problem, we assume that variation across training domains is representative of the variation we might encounter at test time, but also that shifts at test time may be more extreme in magnitude. In particular, we show that reducing differences in risk across training domains can reduce a model's sensitivity to a wide range of extreme distributional shifts, including the challenging setting where the input contains both causal and anti-causal elements. We motivate this approach, Risk Extrapolation (REx), as a form of robust optimization over a perturbation set of extrapolated domains (MM-REx), and propose a penalty on the variance of training risks (V-REx) as a simpler variant. We prove that variants of REx can recover the causal mechanisms of the targets, while also providing some robustness to changes in the input distribution ("covariate shift"). By appropriately trading-off robustness to causally induced distributional shifts and covariate shift, REx is able to outperform alternative methods such as Invariant Risk Minimization in situations where these types of shift co-occur.
翻译:分配变化是将机器学习预测系统从实验室转移到现实世界的主要障碍之一。 解决这个问题,我们假设,不同培训领域的差异代表了我们在测试时可能遇到的差异,但测试时的变化可能更为极端。 特别是,我们表明,降低不同培训领域的风险差异可以降低模型对一系列极端分配变化的敏感性,包括输入包含因果和反因果要素的具有挑战性的设置。 我们鼓励这一方法,即风险外推法(REx),作为对一套外推域(MM-REx)的渗透性优化的一种形式,并提出对培训风险差异(V-REx)的处罚,作为更简单的变式。我们证明,REx的变方可以恢复目标的因果机制,同时也为投入分布的变化(“ Covoliate 变换 ” ) 提供了一些稳健性。 通过适当的交易对因果性分布变化和变量变换的稳健性,REx能够超越替代方法,如在这类变换情况下的内变量风险最小化。